{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,5]],"date-time":"2026-03-05T15:29:56Z","timestamp":1772724596339,"version":"3.50.1"},"reference-count":49,"publisher":"Association for Computing Machinery (ACM)","issue":"1","license":[{"start":{"date-parts":[[2023,1,20]],"date-time":"2023-01-20T00:00:00Z","timestamp":1674172800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key R&D Program of China","doi-asserted-by":"crossref","award":["2020YFB0906000, and 2020YFB0906001"],"award-info":[{"award-number":["2020YFB0906000, and 2020YFB0906001"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Des. Autom. Electron. Syst."],"published-print":{"date-parts":[[2023,1,31]]},"abstract":"<jats:p>Deep Neural Networks (DNNs) have achieved remarkable success in various Artificial Intelligence applications. Quantization is a critical step in DNNs compression and acceleration for deployment. To further boost DNN execution efficiency, many works explore to leverage the input-dependent redundancy with dynamic quantization for different regions. However, the sensitive regions in the feature map are irregularly distributed, which restricts the real speed up for existing accelerators. To this end, we propose an algorithm-architecture co-design, named Structured Dynamic Precision (SDP). Specifically, we propose a quantization scheme in which the high-order bit part and the low-order bit part of data can be masked independently. And a fixed number of term parts are dynamically selected for computation based on the importance of each term in the group. We also present a hardware design to enable the algorithm efficiently with small overheads, whose inference time mainly scales with the precision proportionally. Evaluation experiments on extensive networks demonstrate that compared to the state-of-the-art dynamic quantization accelerator DRQ, our SDP can achieve 29% performance gain and 51% energy reduction for the same level of model accuracy.<\/jats:p>","DOI":"10.1145\/3549535","type":"journal-article","created":{"date-parts":[[2022,7,19]],"date-time":"2022-07-19T12:16:12Z","timestamp":1658232972000},"page":"1-24","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":6,"title":["Structured Dynamic Precision for Deep Neural Networks Quantization"],"prefix":"10.1145","volume":"28","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2295-5433","authenticated-orcid":false,"given":"Kai","family":"Huang","sequence":"first","affiliation":[{"name":"Zhejiang University, Hangzhou, Zhejiang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7525-9672","authenticated-orcid":false,"given":"Bowen","family":"Li","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, Zhejiang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4882-7504","authenticated-orcid":false,"given":"Dongliang","family":"Xiong","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, Zhejiang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8215-166X","authenticated-orcid":false,"given":"Haitian","family":"Jiang","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, Zhejiang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6283-2262","authenticated-orcid":false,"given":"Xiaowen","family":"Jiang","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, Zhejiang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1515-8989","authenticated-orcid":false,"given":"Xiaolang","family":"Yan","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, Zhejiang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0405-6290","authenticated-orcid":false,"given":"Luc","family":"Claesen","sequence":"additional","affiliation":[{"name":"University of Hasselt, Diepenbeek, Belgium"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7229-179X","authenticated-orcid":false,"given":"Dehong","family":"Liu","sequence":"additional","affiliation":[{"name":"China Southern Power Grid Co., Ltd., Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5409-1425","authenticated-orcid":false,"given":"Junjian","family":"Chen","sequence":"additional","affiliation":[{"name":"China Southern Power Grid Co., Ltd., Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0001-5619","authenticated-orcid":false,"given":"Zhili","family":"Liu","sequence":"additional","affiliation":[{"name":"Sec-Chip Technology Co., Ltd., Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2023,1,20]]},"reference":[{"key":"e_1_3_2_2_2","first-page":"662","volume-title":"Proceedings of the 45th ACM\/IEEE Annual International Symposium on Computer Architecture (ISCA\u201918)","author":"Akhlaghi Vahideh","year":"2018","unstructured":"Vahideh Akhlaghi, Amir Yazdanbakhsh, Kambiz Samadi, Rajesh K. Gupta, and Hadi Esmaeilzadeh. 2018. SnaPEA: Predictive early activation for reducing computation in deep convolutional neural networks. In Proceedings of the 45th ACM\/IEEE Annual International Symposium on Computer Architecture (ISCA\u201918). 662\u2013673."},{"key":"e_1_3_2_3_2","doi-asserted-by":"publisher","DOI":"10.1145\/3123939.3123982"},{"key":"e_1_3_2_4_2","first-page":"1","volume-title":"Proceedings of the 43rd ACM\/IEEE Annual International Symposium on Computer Architecture (ISCA\u201916)","author":"Albericio Jorge","year":"2016","unstructured":"Jorge Albericio, Patrick Judd, Tayler H. Hetherington, Tor M. Aamodt, Natalie D. Enright Jerger, and Andreas Moshovos. 2016. Cnvlutin: Ineffectual-neuron-free deep neural network computing. In Proceedings of the 43rd ACM\/IEEE Annual International Symposium on Computer Architecture (ISCA\u201916). 1\u201313."},{"key":"e_1_3_2_5_2","article-title":"Estimating or propagating gradients through stochastic neurons for conditional computation","volume":"1308","author":"Bengio Yoshua","year":"2013","unstructured":"Yoshua Bengio, Nicholas L\u00e9onard, and Aaron C. Courville. 2013. Estimating or propagating gradients through stochastic neurons for conditional computation. CoRR abs\/1308.3432.","journal-title":"CoRR"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1145\/3007787.3001177"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1109\/MICRO.2014.58"},{"key":"e_1_3_2_8_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2020.2976475"},{"key":"e_1_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2015.2439281"},{"key":"e_1_3_2_11_2","first-page":"293","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV\u201919)","author":"Dong Zhen","year":"2019","unstructured":"Zhen Dong, Zhewei Yao, Amir Gholami, Michael W. Mahoney, and Kurt Keutzer. 2019. HAWQ: Hessian AWare quantization of neural networks with mixed-precision. In Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV\u201919). 293\u2013302."},{"key":"e_1_3_2_12_2","volume-title":"Proceedings of the 7th International Conference on Learning Representations (ICLR\u201919)","author":"Gao Xitong","year":"2019","unstructured":"Xitong Gao, Yiren Zhao, Lukasz Dudziak, Robert D. Mullins, and Cheng-Zhong Xu. 2019. Dynamic channel pruning: Feature boosting and suppression. In Proceedings of the 7th International Conference on Learning Representations (ICLR\u201919)."},{"key":"e_1_3_2_13_2","doi-asserted-by":"crossref","first-page":"580","DOI":"10.1109\/CVPR.2014.81","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201914)","author":"Girshick Ross B.","year":"2014","unstructured":"Ross B. Girshick, Jeff Donahue, Trevor Darrell, and Jitendra Malik. 2014. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201914). 580\u2013587."},{"key":"e_1_3_2_14_2","first-page":"4851","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV\u201919)","author":"Gong Ruihao","year":"2019","unstructured":"Ruihao Gong, Xianglong Liu, Shenghu Jiang, Tianxiang Li, Peng Hu, Jiazhen Lin, Fengwei Yu, and Junjie Yan. 2019. Differentiable soft quantization: Bridging full-precision and low-bit neural networks. In Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV\u201919). 4851\u20134860."},{"key":"e_1_3_2_15_2","volume-title":"Proceedings of the 4th International Conference on Learning Representations (ICLR\u201916)","author":"Han Song","year":"2016","unstructured":"Song Han, Huizi Mao, and William J. Dally. 2016. Deep compression: Compressing deep neural network with pruning, trained quantization and huffman coding. In Proceedings of the 4th International Conference on Learning Representations (ICLR\u201916)."},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_17_2","first-page":"4107","volume-title":"Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems","author":"Hubara Itay","year":"2016","unstructured":"Itay Hubara, Matthieu Courbariaux, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio. 2016. Binarized neural networks. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems. 4107\u20134115."},{"key":"e_1_3_2_18_2","first-page":"2704","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201918)","author":"Jacob Benoit","year":"2018","unstructured":"Benoit Jacob, Skirmantas Kligys, Bo Chen, Menglong Zhu, Matthew Tang, Andrew G. Howard, Hartwig Adam, and Dmitry Kalenichenko. 2018. Quantization and training of neural networks for efficient integer-arithmetic-only inference. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201918). 2704\u20132713."},{"key":"e_1_3_2_19_2","first-page":"2704","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201918)","author":"Jacob Benoit","year":"2018","unstructured":"Benoit Jacob, Skirmantas Kligys, Bo Chen, Menglong Zhu, Matthew Tang, Andrew G. Howard, Hartwig Adam, and Dmitry Kalenichenko. 2018. Quantization and training of neural networks for efficient integer-arithmetic-only inference. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201918). 2704\u20132713."},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.1145\/3079856.3080246"},{"key":"e_1_3_2_21_2","first-page":"19:1\u201319:12","volume-title":"Proceedings of the 49th Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO\u201916)","author":"Judd Patrick","year":"2016","unstructured":"Patrick Judd, Jorge Albericio, Tayler H. Hetherington, Tor M. Aamodt, and Andreas Moshovos. 2016. Stripes: Bit-serial deep neural network computing. In Proceedings of the 49th Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO\u201916). 19:1\u201319:12."},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1109\/JETCAS.2018.2865006"},{"key":"e_1_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.1145\/3065386"},{"key":"e_1_3_2_24_2","first-page":"96","volume-title":"Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis (SC\u201920)","author":"Kung Hsiang-Tsung","year":"2020","unstructured":"Hsiang-Tsung Kung, Bradley McDanel, and Sai Qian Zhang. 2020. Term quantization: Furthering quantization at run time. In Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis (SC\u201920). 96."},{"key":"e_1_3_2_25_2","first-page":"821","volume-title":"Proceedings of the 24th International Conference on Architectural Support for Programming Languages and Operating Systems (ASPLOS\u201919)","author":"Kung H. T.","year":"2019","unstructured":"H. T. Kung, Bradley McDanel, and Sai Qian Zhang. 2019. Packing sparse convolutional neural networks for efficient systolic array implementations: Column combining under joint optimization. In Proceedings of the 24th International Conference on Architectural Support for Programming Languages and Operating Systems (ASPLOS\u201919). 821\u2013834."},{"key":"e_1_3_2_26_2","doi-asserted-by":"crossref","first-page":"28","DOI":"10.1145\/3352460.3358295","volume-title":"Proceedings of the 52nd Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO\u201919)","author":"Lascorz Alberto Delmas","year":"2019","unstructured":"Alberto Delmas Lascorz, Sayeh Sharify, Isak Edo Vivancos, Dylan Malone Stuart, Omar Mohamed Awad, Patrick Judd, Mostafa Mahmoud, Milos Nikolic, Kevin Siu, Zissis Poulos, and Andreas Moshovos. 2019. ShapeShifter: Enabling fine-grain data width adaptation in deep learning. In Proceedings of the 52nd Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO\u201919). 28\u201341."},{"key":"e_1_3_2_27_2","unstructured":"Changlin Li Guangrun Wang Bing Wang Xiaodan Liang Zhihui Li and Xiaojun Chang. 2021. Dynamic slimmable network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201921) . 8607\u20138617."},{"key":"e_1_3_2_28_2","article-title":"Ternary weight networks","volume":"1605","author":"Li Fengfu","year":"2016","unstructured":"Fengfu Li and Bin Liu. 2016. Ternary weight networks. CoRR abs\/1605.04711 (2016).","journal-title":"CoRR"},{"key":"e_1_3_2_29_2","first-page":"738","volume-title":"Proceedings of the 53rd Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO\u201920)","author":"Liu Liu","year":"2020","unstructured":"Liu Liu, Zheng Qu, Lei Deng, Fengbin Tu, Shuangchen Li, Xing Hu, Zhenyu Gu, Yufei Ding, and Yuan Xie. 2020. DUET: Boosting deep neural network efficiency on dual-module architecture. In Proceedings of the 53rd Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO\u201920). 738\u2013750."},{"key":"e_1_3_2_30_2","article-title":"NVIDIA\u2019s Automatic SParsity (ASP) Library","year":"2020","unstructured":"NVIDIA. 2020. NVIDIA\u2019s Automatic SParsity (ASP) Library. Retrieved from https:\/\/github.com\/NVIDIA\/apex\/tree\/master\/apex\/contrib\/sparsity.","journal-title":"https:\/\/github.com\/NVIDIA\/apex\/tree\/master\/apex\/contrib\/sparsity"},{"key":"e_1_3_2_31_2","first-page":"8024","volume-title":"Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems (NeurIPS\u201919)","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas K\u00f6pf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems (NeurIPS\u201919). 8024\u20138035."},{"key":"e_1_3_2_32_2","doi-asserted-by":"publisher","DOI":"10.1109\/L-CA.2011.4"},{"key":"e_1_3_2_33_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00474"},{"key":"e_1_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.1145\/3307650.3322255"},{"key":"e_1_3_2_35_2","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA.2018.00069"},{"key":"e_1_3_2_36_2","first-page":"256","volume-title":"Proceedings of the 53rd Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO\u201920)","author":"Shomron Gil","year":"2020","unstructured":"Gil Shomron and Uri C. Weiser. 2020. Non-blocking simultaneous multithreading: Embracing the resiliency of deep neural networks. In Proceedings of the 53rd Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO\u201920). 256\u2013269."},{"key":"e_1_3_2_37_2","volume-title":"Proceedings of the 3rd International Conference on Learning Representations (ICLR\u201915)","author":"Simonyan Karen","year":"2015","unstructured":"Karen Simonyan and Andrew Zisserman. 2015. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR\u201915)."},{"key":"e_1_3_2_38_2","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA.2018.00068"},{"key":"e_1_3_2_39_2","first-page":"1010","volume-title":"Proceedings of the 47th ACM\/IEEE Annual International Symposium on Computer Architecture (ISCA\u201920)","author":"Song Zhuoran","year":"2020","unstructured":"Zhuoran Song, Bangqi Fu, Feiyang Wu, Zhaoming Jiang, Li Jiang, Naifeng Jing, and Xiaoyao Liang. 2020. DRQ: Dynamic region-based quantization for deep neural network acceleration. In Proceedings of the 47th ACM\/IEEE Annual International Symposium on Computer Architecture (ISCA\u201920). 1010\u20131021."},{"key":"e_1_3_2_40_2","first-page":"29","volume-title":"Proceedings of the 14th International Symposium on Applied Reconfigurable Computing: Architectures, Tools, and Applications ARC\u201918)","author":"Su Jiang","year":"2018","unstructured":"Jiang Su, Nicholas J. Fraser, Giulio Gambardella, Michaela Blott, Gianluca Durelli, David B. Thomas, Philip Heng Wai Leong, and Peter Y. K. Cheung. 2018. Accuracy to throughput trade-offs for reduced precision neural networks on reconfigurable logic. In Proceedings of the 14th International Symposium on Applied Reconfigurable Computing: Architectures, Tools, and Applications ARC\u201918). 29\u201342."},{"key":"e_1_3_2_41_2","first-page":"307","volume-title":"Proceedings of the 28th International Conference on Field Programmable Logic and Applications (FPL\u201918)","author":"Umuroglu Yaman","year":"2018","unstructured":"Yaman Umuroglu, Lahiru Rasnayake, and Magnus Sj\u00e4lander. 2018. BISMO: A scalable bit-serial matrix multiplication overlay for reconfigurable computing. In Proceedings of the 28th International Conference on Field Programmable Logic and Applications (FPL\u201918). 307\u2013314."},{"key":"e_1_3_2_42_2","first-page":"97","volume-title":"Proceedings of the IEEE International Symposium on High-Performance Computer Architecture (HPCA\u201921)","author":"Wang Hanrui","year":"2021","unstructured":"Hanrui Wang, Zhekai Zhang, and Song Han. 2021. SpAtten: Efficient sparse attention architecture with cascade token and head pruning. In Proceedings of the IEEE International Symposium on High-Performance Computer Architecture (HPCA\u201921). 97\u2013110."},{"key":"e_1_3_2_43_2","first-page":"8612","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201919)","author":"Wang Kuan","year":"2019","unstructured":"Kuan Wang, Zhijian Liu, Yujun Lin, Ji Lin, and Song Han. 2019. HAQ: Hardware-aware automated quantization with mixed precision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201919). 8612\u20138620."},{"key":"e_1_3_2_44_2","first-page":"2074","volume-title":"Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems","author":"Wen Wei","year":"2016","unstructured":"Wei Wen, Chunpeng Wu, Yandan Wang, Yiran Chen, and Hai Li. 2016. Learning structured sparsity in deep neural networks. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems. 2074\u20132082."},{"key":"e_1_3_2_45_2","first-page":"11875","volume-title":"Proceedings of the 38th International Conference on Machine Learning (ICML\u201921)","author":"Yao Zhewei","year":"2021","unstructured":"Zhewei Yao, Zhen Dong, Zhangcheng Zheng, Amir Gholami, Jiali Yu, Eric Tan, Leyuan Wang, Qijing Huang, Yida Wang, Michael W. Mahoney, and Kurt Keutzer. 2021. HAWQ-V3: Dyadic neural network quantization. In Proceedings of the 38th International Conference on Machine Learning (ICML\u201921). 11875\u201311886."},{"key":"e_1_3_2_46_2","first-page":"548","volume-title":"Proceedings of the 44th Annual International Symposium on Computer Architecture (ISCA\u201917)","author":"Yu Jiecao","year":"2017","unstructured":"Jiecao Yu, Andrew Lukefahr, David J. Palframan, Ganesh S. Dasika, Reetuparna Das, and Scott A. Mahlke. 2017. Scalpel: Customizing DNN pruning to the underlying hardware parallelism. In Proceedings of the 44th Annual International Symposium on Computer Architecture (ISCA\u201917). 548\u2013560."},{"key":"e_1_3_2_47_2","first-page":"20:1\u201320:12","volume-title":"Proceedings of the 49th Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO\u201916)","author":"Zhang Shijin","year":"2016","unstructured":"Shijin Zhang, Zidong Du, Lei Zhang, Huiying Lan, Shaoli Liu, Ling Li, Qi Guo, Tianshi Chen, and Yunji Chen. 2016. Cambricon-X: An accelerator for sparse neural networks. In Proceedings of the 49th Annual IEEE\/ACM International Symposium on Microarchitecture (MICRO\u201916). 20:1\u201320:12."},{"key":"e_1_3_2_48_2","volume-title":"Proceedings of the 8th International Conference on Learning Representations (ICLR\u201920)","author":"Zhang Yichi","year":"2020","unstructured":"Yichi Zhang, Ritchie Zhao, Weizhe Hua, Nayun Xu, G. Edward Suh, and Zhiru Zhang. 2020. Precision gating: Improving neural network efficiency with dynamic dual-precision activations. In Proceedings of the 8th International Conference on Learning Representations (ICLR\u201920)."},{"key":"e_1_3_2_49_2","volume-title":"Proceedings of the 5th International Conference on Learning Representations (ICLR\u201917)","author":"Zhou Aojun","year":"2017","unstructured":"Aojun Zhou, Anbang Yao, Yiwen Guo, Lin Xu, and Yurong Chen. 2017. Incremental network quantization: Towards lossless CNNs with low-precision weights. In Proceedings of the 5th International Conference on Learning Representations (ICLR\u201917)."},{"key":"e_1_3_2_50_2","article-title":"DoReFa-Net: Training low bitwidth convolutional neural networks with low bitwidth gradients","volume":"1606","author":"Zhou Shuchang","year":"2016","unstructured":"Shuchang Zhou, Zekun Ni, Xinyu Zhou, He Wen, Yuxin Wu, and Yuheng Zou. 2016. DoReFa-Net: Training low bitwidth convolutional neural networks with low bitwidth gradients. CoRR abs\/1606.06160.","journal-title":"CoRR"}],"container-title":["ACM Transactions on Design Automation of Electronic Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3549535","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3549535","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:00:11Z","timestamp":1750186811000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3549535"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,1,20]]},"references-count":49,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2023,1,31]]}},"alternative-id":["10.1145\/3549535"],"URL":"https:\/\/doi.org\/10.1145\/3549535","relation":{},"ISSN":["1084-4309","1557-7309"],"issn-type":[{"value":"1084-4309","type":"print"},{"value":"1557-7309","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,1,20]]},"assertion":[{"value":"2021-11-22","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2022-06-20","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-01-20","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}