{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,21]],"date-time":"2025-11-21T05:52:00Z","timestamp":1763704320586,"version":"3.45.0"},"reference-count":43,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,10,26]],"date-time":"2025-10-26T00:00:00Z","timestamp":1761436800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,10,26]],"date-time":"2025-10-26T00:00:00Z","timestamp":1761436800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100006190","name":"Research and Development","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100006190","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100007219","name":"Natural Science Foundation of Shanghai","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100007219","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,10,26]]},"DOI":"10.1109\/iccad66269.2025.11240736","type":"proceedings-article","created":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T18:39:34Z","timestamp":1763663974000},"page":"1-9","source":"Crossref","is-referenced-by-count":0,"title":["QUARK: Quantization-Enabled Circuit Sharing for Transformer Acceleration by Exploiting Common Patterns in Nonlinear Operations"],"prefix":"10.1109","author":[{"given":"Zhixiong","family":"Zhao","sequence":"first","affiliation":[{"name":"Nanyang Technological University"}]},{"given":"Haomin","family":"Li","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University,School of Computer Science"}]},{"given":"Fangxin","family":"Liu","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University,School of Computer Science"}]},{"given":"Yuncheng","family":"Lu","sequence":"additional","affiliation":[{"name":"Nanyang Technological University"}]},{"given":"Zongwu","family":"Wang","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University,School of Computer Science"}]},{"given":"Tao","family":"Yang","sequence":"additional","affiliation":[{"name":"Huawei Technologies Co., Ltd"}]},{"given":"Li","family":"Jiang","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University,School of Computer Science"}]},{"given":"Haibing","family":"Guan","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University,School of Computer Science"}]}],"member":"263","reference":[{"article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","year":"2021","author":"Dosovitskiy","key":"ref1"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00523"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1145\/3140659.3080246"},{"key":"ref4","first-page":"5506","article-title":"I-bert: Integer-only bert quantization","volume-title":"International conference on machine learning","author":"Kim"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3649329.3655896"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA57654.2024.00082"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/DAC18074.2021.9586134"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/3489517.3530505"},{"article-title":"Fq-vit: Post-training quantization for fully quantized vision transformer","year":"2023","author":"Lin","key":"ref9"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICCAD57390.2023.10323725"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01565"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICDSP.2018.8631588"},{"key":"ref13","first-page":"315","article-title":"Deep sparse rectifier neural networks","volume-title":"Proceedings of the fourteenth international conference on artificial intelligence and statistics. JMLR Workshop and Conference Proceedings","author":"Glorot"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4684-9316-0"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA47549.2020.00035"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA51647.2021.00018"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA52012.2021.00060"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TCSII.2021.3120495"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/AICAS59952.2024.10595882"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6409"},{"key":"ref21","first-page":"873","article-title":"Vs-quant: Per-vector scaled quantization for accurate low-precision neural network inference","volume-title":"Proceedings of Machine Learning and Systems","volume":"3","author":"Dai"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.627"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2022.3229313"},{"article-title":"Gaussian error linear units (gelus)","year":"2016","author":"Hendrycks","key":"ref24"},{"article-title":"Layer normalization","year":"2016","author":"Lei Ba","key":"ref25"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/164"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19775-8_12"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01580"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/emc2-nips53020.2019.00009"},{"article-title":"Easyquant: Post-training quantization via scale optimization","year":"2020","author":"Wu","key":"ref30"},{"key":"ref31","first-page":"17 402","article-title":"Outlier suppression: Pushing the limit of low-bit transformer language models","volume":"35","author":"Wei","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref33","first-page":"10 347","article-title":"Training data-efficient image transformers & distillation through attention","volume-title":"International conference on machine learning","author":"Touvron"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W18-5446"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/n19\u20131423"},{"article-title":"Roberta: A robustly optimized bert pretraining approach","year":"2019","author":"Liu","key":"ref37"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/TCAD.2022.3197489"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1145\/3489517.3530618"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/TCSI.2023.3312775"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/TCAD.2024.3487781"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/ASP-DAC58780.2024.10473931"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/SOCC49529.2020.9524802"}],"event":{"name":"2025 IEEE\/ACM International Conference On Computer Aided Design (ICCAD)","start":{"date-parts":[[2025,10,26]]},"location":"Munich, Germany","end":{"date-parts":[[2025,10,30]]}},"container-title":["2025 IEEE\/ACM International Conference On Computer Aided Design (ICCAD)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11240608\/11240621\/11240736.pdf?arnumber=11240736","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,21]],"date-time":"2025-11-21T05:43:39Z","timestamp":1763703819000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11240736\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,26]]},"references-count":43,"URL":"https:\/\/doi.org\/10.1109\/iccad66269.2025.11240736","relation":{},"subject":[],"published":{"date-parts":[[2025,10,26]]}}}