{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T10:09:07Z","timestamp":1764842947731,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":17,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,6,23]],"date-time":"2024-06-23T00:00:00Z","timestamp":1719100800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2023YFB4404603","2020YFB2206002"],"award-info":[{"award-number":["2023YFB4404603","2020YFB2206002"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["92164301","62204003","62225401","61927901"],"award-info":[{"award-number":["92164301","62204003","62225401","61927901"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005089","name":"Beijing Municipal Natural Science Foundation","doi-asserted-by":"publisher","award":["4232067"],"award-info":[{"award-number":["4232067"]}],"id":[{"id":"10.13039\/501100005089","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,6,23]]},"DOI":"10.1145\/3649329.3658244","type":"proceedings-article","created":{"date-parts":[[2024,11,7]],"date-time":"2024-11-07T19:27:22Z","timestamp":1731007642000},"page":"1-6","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["An In-Memory Computing Accelerator with Reconfigurable Dataflow for Multi-Scale Vision Transformer with Hybrid Topology"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-9866-068X","authenticated-orcid":false,"given":"Zhiyuan","family":"Chen","sequence":"first","affiliation":[{"name":"Peking University, Beijing, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2670-524X","authenticated-orcid":false,"given":"Yufei","family":"Ma","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-6307-7306","authenticated-orcid":false,"given":"Keyi","family":"Li","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-2802-2453","authenticated-orcid":false,"given":"Yifan","family":"Jia","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-6685-8997","authenticated-orcid":false,"given":"Guoxiang","family":"Li","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7676-343X","authenticated-orcid":false,"given":"Meng","family":"Wu","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4570-4613","authenticated-orcid":false,"given":"Tianyu","family":"Jia","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0599-7762","authenticated-orcid":false,"given":"Le","family":"Ye","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8146-4821","authenticated-orcid":false,"given":"Ru","family":"Huang","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, N\/A, China"}]}],"member":"320","published-online":{"date-parts":[[2024,11,7]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Advances in Neural Information Processing Systems","volume":"30","author":"Ashish","year":"2017","unstructured":"Ashish Vaswani et al. Attention is all you need. In Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017."},{"key":"e_1_3_2_1_2_1","volume-title":"International Conference on Learning Representations","author":"Alexey","year":"2021","unstructured":"Alexey Dosovitskiy et al. An image is worth 16\u00d716 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2021."},{"key":"e_1_3_2_1_3_1","first-page":"22","volume-title":"2021 IEEE\/CVF International Conference on Computer Vision (ICCV)","author":"Haiping","year":"2021","unstructured":"Haiping Wu et al. Cvt: Introducing convolutions to vision transformers. In 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pages 22--31, 2021."},{"key":"e_1_3_2_1_4_1","volume-title":"International Conference on Learning Representations","author":"Mehta Sachin","year":"2022","unstructured":"Sachin Mehta and Mohammad Rastegari. Mobilevit: Light-weight, generalpurpose, and mobile-friendly vision transformer. In International Conference on Learning Representations, 2022."},{"key":"e_1_3_2_1_5_1","first-page":"252","volume-title":"2021 IEEE International Solid-State Circuits Conference (ISSCC)","volume":"64","author":"Yu-Der","year":"2021","unstructured":"Yu-Der Chih et al. 16.4 an 89tops\/w and 16.3tops\/mm2 all-digital sram-based full-precision compute-in memory macro in 22nm for machine-learning edge applications. In 2021 IEEE International Solid-State Circuits Conference (ISSCC), volume 64, pages 252--254, 2021."},{"key":"e_1_3_2_1_6_1","first-page":"692","volume-title":"2021 ACM\/IEEE 48th Annual International Symposium on Computer Architecture (ISCA)","author":"Jun Tae","year":"2021","unstructured":"Tae Jun Ham et al. Elsa: Hardware-software co-design for efficient, lightweight self-attention mechanism in neural networks. In 2021 ACM\/IEEE 48th Annual International Symposium on Computer Architecture (ISCA), pages 692--705, 2021."},{"key":"e_1_3_2_1_7_1","first-page":"1","volume-title":"2023 60th ACM\/IEEE Design Automation Conference (DAC)","author":"Zhican","year":"2023","unstructured":"Zhican Wang et al. Cosa:co-operative systolic arrays for multi-head attention mechanism in neural network using hybrid data reuse and fusion methodologies. In 2023 60th ACM\/IEEE Design Automation Conference (DAC), pages 1--6, 2023."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/JSSC.2022.3213542"},{"key":"e_1_3_2_1_9_1","first-page":"9992","volume-title":"2021 IEEE\/CVF International Conference on Computer Vision (ICCV)","author":"Ze","year":"2021","unstructured":"Ze Liu et al. Swin transformer: Hierarchical vision transformer using shifted windows. In 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pages 9992--10002, 2021."},{"key":"e_1_3_2_1_10_1","first-page":"548","volume-title":"2021 IEEE\/CVF International Conference on Computer Vision (ICCV)","author":"Wenhai","year":"2021","unstructured":"Wenhai Wang et al. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pages 548--558, 2021."},{"key":"e_1_3_2_1_11_1","first-page":"1","volume-title":"Automation Test in Europe Conference Exhibition (DATE)","author":"Wantong","year":"2023","unstructured":"Wantong Li et al. Rawatten: Reconfigurable accelerator for window attention in hierarchical vision transformers. In 2023 Design, Automation Test in Europe Conference Exhibition (DATE), pages 1--6, 2023."},{"key":"e_1_3_2_1_12_1","volume-title":"Mnsim 2.0: A behavior-level modeling tool for processing-in-memory architectures","author":"Zhenhua Zhu","year":"2023","unstructured":"Zhenhua Zhu et al. Mnsim 2.0: A behavior-level modeling tool for processing-in-memory architectures. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 42(11):4112--4125, 2023."},{"key":"e_1_3_2_1_13_1","first-page":"16344","volume-title":"Advances in Neural Information Processing Systems","volume":"35","author":"Tri","year":"2022","unstructured":"Tri Dao et al. Flashattention: Fast and memory-efficient exact attention with io-awareness. In S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, editors, Advances in Neural Information Processing Systems, volume 35, pages 16344--16359. Curran Associates, Inc., 2022."},{"key":"e_1_3_2_1_14_1","first-page":"577","volume-title":"Proceedings of the 59th ACM\/IEEE Design Automation Conference, DAC '22","author":"Joonsang","year":"2022","unstructured":"Joonsang Yu et al. Nn-lut: Neural approximation of non-linear operations for efficient transformer inference. In Proceedings of the 59th ACM\/IEEE Design Automation Conference, DAC '22, page 577--582, New York, NY, USA, 2022."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3085572"},{"key":"e_1_3_2_1_16_1","first-page":"10347","volume-title":"Proceedings of the 38th International Conference on Machine Learning","volume":"139","author":"Hugo","year":"2021","unstructured":"Hugo Touvron et al. Training data-efficient image transformers amp; distillation through attention. In Proceedings of the 38th International Conference on Machine Learning, volume 139, pages 10347--10357, 18--24 Jul 2021."},{"key":"e_1_3_2_1_17_1","first-page":"97","volume-title":"2021 IEEE International Symposium on High-Performance Computer Architecture (HPCA)","author":"Hanrui","year":"2021","unstructured":"Hanrui Wang et al. Spatten: Efficient sparse attention architecture with cascade token and head pruning. In 2021 IEEE International Symposium on High-Performance Computer Architecture (HPCA), pages 97--110, 2021."}],"event":{"name":"DAC '24: 61st ACM\/IEEE Design Automation Conference","sponsor":["SIGDA ACM Special Interest Group on Design Automation","IEEE-CEDA","SIGBED ACM Special Interest Group on Embedded Systems"],"location":"San Francisco CA USA","acronym":"DAC '24"},"container-title":["Proceedings of the 61st ACM\/IEEE Design Automation Conference"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3649329.3658244","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3649329.3658244","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:01Z","timestamp":1750295881000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3649329.3658244"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,23]]},"references-count":17,"alternative-id":["10.1145\/3649329.3658244","10.1145\/3649329"],"URL":"https:\/\/doi.org\/10.1145\/3649329.3658244","relation":{},"subject":[],"published":{"date-parts":[[2024,6,23]]},"assertion":[{"value":"2024-11-07","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}