{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T11:43:05Z","timestamp":1774611785067,"version":"3.50.1"},"reference-count":24,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"National Key Research and Development Program of China","award":["2022YFE0116700"],"award-info":[{"award-number":["2022YFE0116700"]}]},{"name":"National Key Research and Development Program of China","award":["2021ZD0113202"],"award-info":[{"award-number":["2021ZD0113202"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62171125"],"award-info":[{"award-number":["62171125"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61876037"],"award-info":[{"award-number":["61876037"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Innovation Project of Jiangsu Province","award":["BZ2023042"],"award-info":[{"award-number":["BZ2023042"]}]},{"name":"Innovation Project of Jiangsu Province","award":["BY2022564"],"award-info":[{"award-number":["BY2022564"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Signal Process. Lett."],"published-print":{"date-parts":[[2024]]},"DOI":"10.1109\/lsp.2024.3350811","type":"journal-article","created":{"date-parts":[[2024,1,12]],"date-time":"2024-01-12T19:04:55Z","timestamp":1705086295000},"page":"446-450","source":"Crossref","is-referenced-by-count":2,"title":["Spatial-Enhanced Multi-Level Wavelet Patching in Vision Transformers"],"prefix":"10.1109","volume":"31","author":[{"ORCID":"https:\/\/orcid.org\/0009-0000-7310-9670","authenticated-orcid":false,"given":"Fuzhi","family":"Wu","sequence":"first","affiliation":[{"name":"LIST, Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications, Southeast University, Ministry of Education, Nanjing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7171-1318","authenticated-orcid":false,"given":"Jiasong","family":"Wu","sequence":"additional","affiliation":[{"name":"LIST, Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications, Southeast University, Ministry of Education, Nanjing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3833-7915","authenticated-orcid":false,"given":"Huazhong","family":"Shu","sequence":"additional","affiliation":[{"name":"LIST, Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications, Southeast University, Ministry of Education, Nanjing, China"}]},{"given":"Guy","family":"Carrault","sequence":"additional","affiliation":[{"name":"Univ Rennes, Inserm, LTSI - UMR 1099, Rennes, France"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9434-6341","authenticated-orcid":false,"given":"Lotfi","family":"Senhadji","sequence":"additional","affiliation":[{"name":"Univ Rennes, Inserm, LTSI - UMR 1099, Rennes, France"}]}],"member":"263","reference":[{"key":"ref1","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Dosovitskiy"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2022.3199145"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2022.3175055"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2023.117032"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2022.3172617"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/tnnls.2022.3227717"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00060"},{"key":"ref8","first-page":"15908","article-title":"Transformer in transformer","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Han"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3206108"},{"key":"ref10","article-title":"BOAT: Bilateral local attention vision transformer","author":"Yu","year":"2022"},{"key":"ref11","article-title":"Twins: Revisiting the design of spatial attention in vision transformers","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Chu"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01358"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2023.109532"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19806-9_19"},{"key":"ref15","article-title":"WaveFormer: Linear-time attention with forward and backward wavelet transform","author":"Zhuang","year":"2022"},{"key":"ref16","article-title":"Adaptive wavelet transformer network for 3D shape representation learning","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Huang"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2021.3088052"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2022.109815"},{"key":"ref19","article-title":"SOSR: Source-free image super-resolution with wavelet augmentation transformer","author":"Ai","year":"2023"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.5555\/3454287.3455008"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref23","article-title":"PyramidTNT: Improved transformer-in-transformer baselines with pyramid architecture","author":"Han","year":"2022"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2023.3268446"}],"container-title":["IEEE Signal Processing Letters"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/97\/10380231\/10399340.pdf?arnumber=10399340","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,6]],"date-time":"2024-02-06T20:41:18Z","timestamp":1707252078000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10399340\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"references-count":24,"URL":"https:\/\/doi.org\/10.1109\/lsp.2024.3350811","relation":{},"ISSN":["1070-9908","1558-2361"],"issn-type":[{"value":"1070-9908","type":"print"},{"value":"1558-2361","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]}}}