{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T15:19:54Z","timestamp":1774365594728,"version":"3.50.1"},"reference-count":53,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,2,1]],"date-time":"2023-02-01T00:00:00Z","timestamp":1675209600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,2,1]],"date-time":"2023-02-01T00:00:00Z","timestamp":1675209600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,2]]},"DOI":"10.1109\/hpca56546.2023.10071027","type":"proceedings-article","created":{"date-parts":[[2023,3,24]],"date-time":"2023-03-24T17:42:55Z","timestamp":1679679775000},"page":"273-286","source":"Crossref","is-referenced-by-count":105,"title":["ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design"],"prefix":"10.1109","author":[{"given":"Haoran","family":"You","sequence":"first","affiliation":[{"name":"Georgia Institute of Technology,Atlanta,GA"}]},{"given":"Zhanyi","family":"Sun","sequence":"additional","affiliation":[{"name":"Rice University,Houston,TX"}]},{"given":"Huihong","family":"Shi","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology,Atlanta,GA"}]},{"given":"Zhongzhi","family":"Yu","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology,Atlanta,GA"}]},{"given":"Yang","family":"Zhao","sequence":"additional","affiliation":[{"name":"Rice University,Houston,TX"}]},{"given":"Yongan","family":"Zhang","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology,Atlanta,GA"}]},{"given":"Chaojian","family":"Li","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology,Atlanta,GA"}]},{"given":"Baopu","family":"Li","sequence":"additional","affiliation":[{"name":"Oracle Health and AI,Redwood,CA"}]},{"given":"Yingyan","family":"Lin","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology,Atlanta,GA"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Synopsys design compiler"},{"key":"ref2","article-title":"Synopsys ic compiler ii"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00338"},{"key":"ref4","article-title":"Longformer: The long-document transformer","author":"Beltagy","year":"2020"},{"key":"ref5","first-page":"864","article-title":"Low-rank bottleneck in multi-head attention models","volume-title":"International Conference on Machine Learning","author":"Bhojanapalli"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00041"},{"key":"ref7","first-page":"1691","article-title":"Generative pretraining from pixels","volume-title":"Proceedings of the 37th International Conference on Machine Learning","volume":"119","author":"Chen"},{"key":"ref8","article-title":"Fbnetv3: Joint architecture-recipe search using neural acquisition function","volume":"abs\/2006.02049","author":"Dai","year":"2020","journal-title":"ArXiv"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref10","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","volume-title":"International Conference on Learning Representations","author":"Dosovitskiy"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01204"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA47549.2020.00035"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA52012.2021.00060"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1145\/3352460.3358275"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01172"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00140"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2013.248"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW53098.2021.00342"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW53098.2021.00342"},{"key":"ref21","article-title":"Reformer: The efficient transformer","author":"Kitaev","year":"2020"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2022.3141231"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1145\/3466752.3480125"},{"key":"ref25","article-title":"Mobilevit: light-weight, general-purpose, and mobile-friendly vision transformer","author":"Mehta","year":"2021"},{"key":"ref26","article-title":"NVIDIA Jetson TX2"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA.2018.00067"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA47549.2020.00015"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.232"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/3503222.3507738"},{"key":"ref31","article-title":"Language models are unsupervised multitask learners","author":"Radford","year":"2019"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1145\/3079856.3080221"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/MICRO50266.2020.00068"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/MICRO50266.2020.00068"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA47549.2020.00062"},{"key":"ref36","article-title":"Vaqf: Fully automatic software-hardware co-design framework for low-bit vision transformer","author":"Sun","year":"2022"},{"key":"ref37","article-title":"Training data-efficient image transformers & distillation through attention","author":"Touvron","year":"2020"},{"key":"ref38","first-page":"10 347","article-title":"Training data-efficient image transformers & distillation through attention","volume-title":"Proceedings of the 38th International Conference on Machine Learning","volume":"139","author":"Touvron"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.spnlp-1.7"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref41","article-title":"Attention is all you need","author":"Vaswani","year":"2017"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA51647.2021.00018"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/iccv48922.2021.00064"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00009"},{"key":"ref46","article-title":"Lite transformer with long-short range attention","volume-title":"International Conference on Learning Representations","author":"Wu"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr46437.2021.00382"},{"key":"ref48","article-title":"Big bird: Transformers for longer sequences","author":"Zaheer","year":"2020","journal-title":"NeurIPS"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1145\/3445814.3446702"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/cvprw56347.2022.00309"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1145\/3240765.3240801"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA47549.2020.00030"},{"key":"ref53","article-title":"Long-short transformer: Efficient transformers for language and vision","volume":"34","author":"Zhu","year":"2021","journal-title":"Advances in Neural Information Processing Systems"}],"event":{"name":"2023 IEEE International Symposium on High-Performance Computer Architecture (HPCA)","location":"Montreal, QC, Canada","start":{"date-parts":[[2023,2,25]]},"end":{"date-parts":[[2023,3,1]]}},"container-title":["2023 IEEE International Symposium on High-Performance Computer Architecture (HPCA)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10070856\/10070923\/10071027.pdf?arnumber=10071027","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,13]],"date-time":"2024-02-13T13:21:36Z","timestamp":1707830496000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10071027\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,2]]},"references-count":53,"URL":"https:\/\/doi.org\/10.1109\/hpca56546.2023.10071027","relation":{},"subject":[],"published":{"date-parts":[[2023,2]]}}}