{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T21:30:31Z","timestamp":1773264631447,"version":"3.50.1"},"reference-count":22,"publisher":"IEEE","license":[{"start":{"date-parts":[[2026,1,19]],"date-time":"2026-01-19T00:00:00Z","timestamp":1768780800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,1,19]],"date-time":"2026-01-19T00:00:00Z","timestamp":1768780800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026,1,19]]},"DOI":"10.1109\/asp-dac66049.2026.11420796","type":"proceedings-article","created":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T19:51:15Z","timestamp":1773172275000},"page":"140-146","source":"Crossref","is-referenced-by-count":0,"title":["PipeViT: Accelerating Vision Transformers via Intra-Layer Pipelining"],"prefix":"10.1109","author":[{"given":"Xilang","family":"Zhou","sequence":"first","affiliation":[{"name":"Fudan University,State Key Lab of Integrated Chips &#x0026; Systems, and School of Microelectronics,Shanghai,China"}]},{"given":"Yiheng","family":"Xu","sequence":"additional","affiliation":[{"name":"Fudan University,State Key Lab of Integrated Chips &#x0026; Systems, and School of Microelectronics,Shanghai,China"}]},{"given":"Haodong","family":"Lu","sequence":"additional","affiliation":[{"name":"Fudan University,State Key Lab of Integrated Chips &#x0026; Systems, and School of Microelectronics,Shanghai,China"}]},{"given":"Jun","family":"Yu","sequence":"additional","affiliation":[{"name":"Fudan University,State Key Lab of Integrated Chips &#x0026; Systems, and School of Microelectronics,Shanghai,China"}]},{"given":"Kun","family":"Wang","sequence":"additional","affiliation":[{"name":"Fudan University,State Key Lab of Integrated Chips &#x0026; Systems, and School of Microelectronics,Shanghai,China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-024-63094-9"},{"key":"ref2","article-title":"Vit-1.58b: Mobile vision transformers in the 1-bit era","author":"Yuan","year":"2024","journal-title":"arXiv preprint arXiv:2406.18051"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.52202\/079017-1272"},{"key":"ref4","article-title":"Heuristical comparison of vision transformers against convolutional neural networks for semantic segmentation on remote sensing imagery","author":"Murad","year":"2024","journal-title":"arXiv preprint arXiv:2411.09101"},{"key":"ref5","article-title":"Consept: Continual semantic segmentation via adapter-based vision transformer","author":"Dong","year":"2024","journal-title":"arXiv preprint arXiv:2402.16674"},{"key":"ref6","article-title":"Interpretable image classification with adaptive prototype-based vision transformers","author":"Ma","year":"2024","journal-title":"arXiv preprint arXiv:2410.20722"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA56546.2023.10071027"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA56546.2023.10071081"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA56546.2023.10071047"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01587"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01386"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.936"},{"key":"ref13","article-title":"Vision mamba: Efficient visual representation learning with bidirectional state space model","author":"Zhu","year":"2024","journal-title":"arXiv preprint arXiv:2401.09417"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.52202\/068431-1189"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/3575693.3575747"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICCAD57390.2023.10323651"},{"key":"ref17","article-title":"An image is worth 16\u00d716 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020","journal-title":"arXiv preprint arXiv:2010.11929"},{"key":"ref18","first-page":"10347","article-title":"Training data-efficient image transformers & distillation through attention","volume-title":"Proceedings of the International Conference on Machine Learning (ICML)","author":"Touvron"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01204"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/FPL57034.2022.00027"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ISCAS46773.2023.10181988"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/HiPC58850.2023.00039"}],"event":{"name":"2026 31st Asia and South Pacific Design Automation Conference (ASP-DAC)","location":"Lantau, Hong Kong","start":{"date-parts":[[2026,1,19]]},"end":{"date-parts":[[2026,1,22]]}},"container-title":["2026 31st Asia and South Pacific Design Automation Conference (ASP-DAC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11420221\/11420229\/11420796.pdf?arnumber=11420796","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T05:29:12Z","timestamp":1773206952000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11420796\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,19]]},"references-count":22,"URL":"https:\/\/doi.org\/10.1109\/asp-dac66049.2026.11420796","relation":{},"subject":[],"published":{"date-parts":[[2026,1,19]]}}}