{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T02:17:21Z","timestamp":1775873841855,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":57,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,2,11]],"date-time":"2022-02-11T00:00:00Z","timestamp":1644537600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"NSERC Discovery Grant","award":["RGPIN-2019-04613, DGECR-2019-00120"],"award-info":[{"award-number":["RGPIN-2019-04613, DGECR-2019-00120"]}]},{"name":"CFI John R. Evans Leaders Fund"},{"DOI":"10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["CCF-1901378"],"award-info":[{"award-number":["CCF-1901378"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Alliance Grant","award":["ALLRP-552042-2020"],"award-info":[{"award-number":["ALLRP-552042-2020"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,2,13]]},"DOI":"10.1145\/3490422.3502364","type":"proceedings-article","created":{"date-parts":[[2022,2,12]],"date-time":"2022-02-12T05:09:21Z","timestamp":1644642561000},"page":"134-145","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":80,"title":["FILM-QNN: Efficient FPGA Acceleration of Deep Neural Networks with Intra-Layer, Mixed-Precision Quantization"],"prefix":"10.1145","author":[{"given":"Mengshu","family":"Sun","sequence":"first","affiliation":[{"name":"Northeastern University, Boston, MA, USA"}]},{"given":"Zhengang","family":"Li","sequence":"additional","affiliation":[{"name":"Northeastern University, Boston, MA, USA"}]},{"given":"Alec","family":"Lu","sequence":"additional","affiliation":[{"name":"Simon Fraser University, Burnaby, Canada"}]},{"given":"Yanyu","family":"Li","sequence":"additional","affiliation":[{"name":"Northeastern University, Boston, MA, USA"}]},{"given":"Sung-En","family":"Chang","sequence":"additional","affiliation":[{"name":"Northeastern University, Boston, MA, USA"}]},{"given":"Xiaolong","family":"Ma","sequence":"additional","affiliation":[{"name":"Northeastern University, Boston, MA, USA"}]},{"given":"Xue","family":"Lin","sequence":"additional","affiliation":[{"name":"Northeastern University, Boston, MA, USA"}]},{"given":"Zhenman","family":"Fang","sequence":"additional","affiliation":[{"name":"Simon Fraser University, Burnaby, Canada"}]}],"member":"320","published-online":{"date-parts":[[2022,2,11]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432","author":"Bengio Yoshua","year":"2013","unstructured":"013)]% bengio2013estimating, Yoshua Bengio , Nicholas L\u00e9onard , and Aaron Courville . 2013. Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432 ( 2013 ). 013)]% bengio2013estimating, Yoshua Bengio, Nicholas L\u00e9onard, and Aaron Courville. 2013. Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432 (2013)."},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1109\/HPCA51647.2021.00027"},{"key":"e_1_3_2_2_3_1","volume-title":"The 2019 International Joint Conference on Neural Networks (IJCNN)","author":"Cheng Gong","year":"2019","unstructured":"019)]% cheng2019uL2Q, Gong Cheng , Lu Ye , Li Tao , Zhang Xiaofan , Hao Cong , Chen Deming , and Chen Yao . 2019 . \u03bcL2Q: An Ultra-Low Loss Quantization Method for DNN . The 2019 International Joint Conference on Neural Networks (IJCNN) (2019), 1--8. 019)]% cheng2019uL2Q, Gong Cheng, Lu Ye, Li Tao, Zhang Xiaofan, Hao Cong, Chen Deming, and Chen Yao. 2019. \u03bcL2Q: An Ultra-Low Loss Quantization Method for DNN. The 2019 International Joint Conference on Neural Networks (IJCNN) (2019), 1--8."},{"key":"e_1_3_2_2_4_1","volume-title":"Vijayalakshmi Srinivasan, and Kailash Gopalakrishnan.","author":"Choi Jungwook","year":"2018","unstructured":"018)]% choi2018pact, Jungwook Choi , Zhuo Wang , Swagath Venkataramani , Pierce I-Jen Chuang , Vijayalakshmi Srinivasan, and Kailash Gopalakrishnan. 2018 . Pact : Parameterized clipping activation for quantized neural networks. arXiv preprint arXiv:1805.06085 (2018). 018)]% choi2018pact, Jungwook Choi, Zhuo Wang, Swagath Venkataramani, Pierce I-Jen Chuang, Vijayalakshmi Srinivasan, and Kailash Gopalakrishnan. 2018. Pact: Parameterized clipping activation for quantized neural networks. arXiv preprint arXiv:1805.06085 (2018)."},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/FCCM.2018.00020"},{"key":"e_1_3_2_2_6_1","volume-title":"Binaryconnect: Training deep neural networks with binary weights during propagations. In Advances in neural information processing systems (NeurIPS) . 3123--3131.","author":"Courbariaux Matthieu","year":"2015","unstructured":"015)]% courbariaux2015binaryconnect, Matthieu Courbariaux , Yoshua Bengio , and Jean-Pierre David . 2015 . Binaryconnect: Training deep neural networks with binary weights during propagations. In Advances in neural information processing systems (NeurIPS) . 3123--3131. 015)]% courbariaux2015binaryconnect, Matthieu Courbariaux, Yoshua Bengio, and Jean-Pierre David. 2015. Binaryconnect: Training deep neural networks with binary weights during propagations. In Advances in neural information processing systems (NeurIPS) . 3123--3131."},{"key":"e_1_3_2_2_7_1","unstructured":"016)]% courbariaux2016binarized Matthieu Courbariaux Itay Hubara Daniel Soudry Ran El-Yaniv and Yoshua Bengio. 2016. Binarized neural networks: Training deep neural networks with weights and activations constrained to  016)]% courbariaux2016binarized Matthieu Courbariaux Itay Hubara Daniel Soudry Ran El-Yaniv and Yoshua Bengio. 2016. Binarized neural networks: Training deep neural networks with weights and activations constrained to"},{"key":"e_1_3_2_2_8_1","volume-title":"arXiv preprint arXiv:1602.02830","year":"2016","unstructured":"1 or-1. arXiv preprint arXiv:1602.02830 ( 2016 ). 1 or-1. arXiv preprint arXiv:1602.02830 (2016)."},{"key":"e_1_3_2_2_9_1","volume-title":"2019 a. HAWQ-V2: Hessian Aware trace-Weighted Quantization of neural networks. arXiv preprint arXiv:1911.03852","author":"Dong Zhen","year":"2019","unstructured":"019a)]% dong2019hawqv2, Zhen Dong , Zhewei Yao , Yaohui Cai , Daiyaan Arfeen , Amir Gholami , Michael W Mahoney , and Kurt Keutzer . 2019 a. HAWQ-V2: Hessian Aware trace-Weighted Quantization of neural networks. arXiv preprint arXiv:1911.03852 ( 2019 ). 019a)]% dong2019hawqv2, Zhen Dong, Zhewei Yao, Yaohui Cai, Daiyaan Arfeen, Amir Gholami, Michael W Mahoney, and Kurt Keutzer. 2019 a. HAWQ-V2: Hessian Aware trace-Weighted Quantization of neural networks. arXiv preprint arXiv:1911.03852 (2019)."},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00038"},{"key":"e_1_3_2_2_11_1","volume-title":"International Conference on Learning Representations (ICLR)","author":"Esser Steven K","year":"2019","unstructured":"019)]% esser2019learned, Steven K Esser , Jeffrey L McKinstry , Deepika Bablani , Rathinakumar Appuswamy , and Dharmendra S Modha . 2019 . Learned step size quantization . International Conference on Learning Representations (ICLR) (2019). 019)]% esser2019learned, Steven K Esser, Jeffrey L McKinstry, Deepika Bablani, Rathinakumar Appuswamy, and Dharmendra S Modha. 2019. Learned step size quantization. International Conference on Learning Representations (ICLR) (2019)."},{"key":"e_1_3_2_2_12_1","volume-title":"https:\/\/pytorch.org\/vision\/stable\/models.html Last accessed","year":"2021","unstructured":"021)]% Torchvision, Facebook. 2021. Torchvision. https:\/\/pytorch.org\/vision\/stable\/models.html Last accessed Sept 12, 2021 . 021)]% Torchvision, Facebook. 2021. Torchvision. https:\/\/pytorch.org\/vision\/stable\/models.html Last accessed Sept 12, 2021."},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00495"},{"key":"e_1_3_2_2_14_1","volume-title":"https:\/\/www.tensorflow.org\/lite Last accessed","year":"2021","unstructured":"021)]% TensorFlow-Lite, Google. 2021. TensorFlow. https:\/\/www.tensorflow.org\/lite Last accessed May 27, 2021 . 021)]% TensorFlow-Lite, Google. 2021. TensorFlow. https:\/\/www.tensorflow.org\/lite Last accessed May 27, 2021."},{"key":"e_1_3_2_2_15_1","unstructured":"021)]% nn-accelerator K. Guo W. Li K. Zhong Z. Zhu S. Zeng S. Han Y. Xie P. Debacker M. Verhelst and Y. Wang. 2021. Neural Network Accelerator Comparison. https:\/\/nicsefc.ee.tsinghua.edu.cn\/projects\/neural-network-accelerator\/.  021)]% nn-accelerator K. Guo W. Li K. Zhong Z. Zhu S. Zeng S. Han Y. Xie P. Debacker M. Verhelst and Y. Wang. 2021. Neural Network Accelerator Comparison. https:\/\/nicsefc.ee.tsinghua.edu.cn\/projects\/neural-network-accelerator\/."},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/TCAD.2017.2705069"},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/FPL.2018.00016"},{"key":"e_1_3_2_2_18_1","volume-title":"International Conference on Learning Representations (ICLR)","author":"Han Song","year":"2016","unstructured":"016)]% han2015deep, Song Han , Huizi Mao , and William J Dally . 2016 . Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding . International Conference on Learning Representations (ICLR) (2016). 016)]% han2015deep, Song Han, Huizi Mao, and William J Dally. 2016. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. International Conference on Learning Representations (ICLR) (2016)."},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01170"},{"key":"e_1_3_2_2_20_1","volume-title":"Intel Arria 10 Native Fixed Point DSP IP Core User Guide. https:\/\/www.intel.com\/content\/dam\/www\/programmable\/us\/en\/pdfs\/literature\/ug\/ug_nfp_dsp.pdf Last accessed","year":"2021","unstructured":", Intel. 2017. Intel Arria 10 Native Fixed Point DSP IP Core User Guide. https:\/\/www.intel.com\/content\/dam\/www\/programmable\/us\/en\/pdfs\/literature\/ug\/ug_nfp_dsp.pdf Last accessed Sept 11, 2021 . , Intel. 2017. Intel Arria 10 Native Fixed Point DSP IP Core User Guide. https:\/\/www.intel.com\/content\/dam\/www\/programmable\/us\/en\/pdfs\/literature\/ug\/ug_nfp_dsp.pdf Last accessed Sept 11, 2021."},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.23919\/FPL.2017.8056820"},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/MICRO.2016.7783722"},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00448"},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11713"},{"key":"e_1_3_2_2_25_1","volume-title":"Ternary weight networks. arXiv preprint arXiv:1605.04711","author":"Li Fengfu","year":"2016","unstructured":"016)]% li2016ternary, Fengfu Li , Bo Zhang , and Bin Liu . 2016. Ternary weight networks. arXiv preprint arXiv:1605.04711 ( 2016 ). 016)]% li2016ternary, Fengfu Li, Bo Zhang, and Bin Liu. 2016. Ternary weight networks. arXiv preprint arXiv:1605.04711 (2016)."},{"key":"e_1_3_2_2_26_1","unstructured":"017)]% lin2017towards Xiaofan Lin Cong Zhao and Wei Pan. 2017. Towards accurate binary convolutional neural network. In Advances in Neural Information Processing Systems (NeurIPS). 345--353.  017)]% lin2017towards Xiaofan Lin Cong Zhao and Wei Pan. 2017. Towards accurate binary convolutional neural network. In Advances in Neural Information Processing Systems (NeurIPS). 345--353."},{"key":"e_1_3_2_2_27_1","volume-title":"AutoQ: Automated Kernel-Wise Neural Network Quantization. In International Conference on Learning Representations (ICLR) .","author":"Lou Qian","year":"2019","unstructured":"019)]% lou2019autoq, Qian Lou , Feng Guo , Minje Kim , Lantao Liu , and Lei Jiang . 2019 . AutoQ: Automated Kernel-Wise Neural Network Quantization. In International Conference on Learning Representations (ICLR) . 019)]% lou2019autoq, Qian Lou, Feng Guo, Minje Kim, Lantao Liu, and Lei Jiang. 2019. AutoQ: Automated Kernel-Wise Neural Network Quantization. In International Conference on Learning Representations (ICLR) ."},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3431920.3439284"},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1587\/transinf.2018RCP0008"},{"key":"e_1_3_2_2_30_1","volume-title":"Convolutional neural networks using logarithmic data representation. arXiv preprint arXiv:1603.01025","author":"Miyashita Daisuke","year":"2016","unstructured":"016)]% miyashita2016convolutional, Daisuke Miyashita , Edward H Lee , and Boris Murmann . 2016. Convolutional neural networks using logarithmic data representation. arXiv preprint arXiv:1603.01025 ( 2016 ). 016)]% miyashita2016convolutional, Daisuke Miyashita, Edward H Lee, and Boris Murmann. 2016. Convolutional neural networks using logarithmic data representation. arXiv preprint arXiv:1603.01025 (2016)."},{"key":"e_1_3_2_2_31_1","volume-title":"2017 27th International Conference on Field Programmable Logic and Applications (FPL). IEEE, 1--4.","author":"Nakahara Hiroki","year":"2017","unstructured":"017)]% nakahara2017fully, Hiroki Nakahara , Tomoya Fujii , and Shimpei Sato . 2017 . A fully connected layer elimination for a binarized convolutional neural network on an FPGA . In 2017 27th International Conference on Field Programmable Logic and Applications (FPL). IEEE, 1--4. 017)]% nakahara2017fully, Hiroki Nakahara, Tomoya Fujii, and Shimpei Sato. 2017. A fully connected layer elimination for a binarized convolutional neural network on an FPGA. In 2017 27th International Conference on Field Programmable Logic and Applications (FPL). IEEE, 1--4."},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/FPT.2016.7929552"},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.23919\/DATE.2017.7927113"},{"key":"e_1_3_2_2_34_1","volume-title":"Nvidia Deep Learning Examples. https:\/\/github.com\/NVIDIA\/DeepLearningExamples Last accessed","year":"2021","unstructured":"021)]% Nvidia, Nvidia. 2021. Nvidia Deep Learning Examples. https:\/\/github.com\/NVIDIA\/DeepLearningExamples Last accessed Sept 12, 2021 . 021)]% Nvidia, Nvidia. 2021. Nvidia Deep Learning Examples. https:\/\/github.com\/NVIDIA\/DeepLearningExamples Last accessed Sept 12, 2021."},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01225-0_36"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.23919\/DATE.2017.7927162"},{"key":"e_1_3_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46493-0_32"},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3297858.3304076"},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISCA.2018.00069"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6409"},{"key":"e_1_3_2_2_41_1","volume-title":"International Conference on Learning Representations (ICLR)","author":"Uhlich Stefan","year":"2020","unstructured":"020)]% uhlich2019mixed, Stefan Uhlich , Lukas Mauch , Fabien Cardinaux , Kazuki Yoshiyama , Javier Alonso Garcia , Stephen Tiedemann , Thomas Kemp , and Akira Nakamura . 2020 . Mixed Precision DNNs: All you need is a good parametrization . International Conference on Learning Representations (ICLR) (2020). 020)]% uhlich2019mixed, Stefan Uhlich, Lukas Mauch, Fabien Cardinaux, Kazuki Yoshiyama, Javier Alonso Garcia, Stephen Tiedemann, Thomas Kemp, and Akira Nakamura. 2020. Mixed Precision DNNs: All you need is a good parametrization. International Conference on Learning Representations (ICLR) (2020)."},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3020078.3021744"},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.23919\/FPL.2017.8056863"},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/FPL.2018.00035"},{"key":"e_1_3_2_2_45_1","volume-title":"HAQ: Hardware-Aware Automated Quantization with Mixed Precision. International Conference on Computer Vision and Pattern Recognition (CVPR)","author":"Wang Kuan","year":"2019","unstructured":"019)]% wang2019haq, Kuan Wang , Zhijian Liu , Yujun Lin , Ji Lin , and Song Han . 2019 . HAQ: Hardware-Aware Automated Quantization with Mixed Precision. International Conference on Computer Vision and Pattern Recognition (CVPR) (2019), 8604--8612. 019)]% wang2019haq, Kuan Wang, Zhijian Liu, Yujun Lin, Ji Lin, and Song Han. 2019. HAQ: Hardware-Aware Automated Quantization with Mixed Precision. International Conference on Computer Vision and Pattern Recognition (CVPR) (2019), 8604--8612."},{"key":"e_1_3_2_2_46_1","volume-title":"Mixed precision quantization of convnets via differentiable neural architecture search. arXiv preprint arXiv:1812.00090","author":"Wu Bichen","year":"2018","unstructured":"018)]% wu2018mixed, Bichen Wu , Yanghan Wang , Peizhao Zhang , Yuandong Tian , Peter Vajda , and Kurt Keutzer . 2018. Mixed precision quantization of convnets via differentiable neural architecture search. arXiv preprint arXiv:1812.00090 ( 2018 ). 018)]% wu2018mixed, Bichen Wu, Yanghan Wang, Peizhao Zhang, Yuandong Tian, Peter Vajda, and Kurt Keutzer. 2018. Mixed precision quantization of convnets via differentiable neural architecture search. arXiv preprint arXiv:1812.00090 (2018)."},{"key":"e_1_3_2_2_47_1","volume-title":"Deep Learning with INT8 Optimization on Xilinx Devices. https:\/\/www.xilinx.com\/support\/documentation\/white_papers\/wp486-deep-learning-int8.pdf Last accessed","year":"2021","unstructured":", Xilinx. 2017. Deep Learning with INT8 Optimization on Xilinx Devices. https:\/\/www.xilinx.com\/support\/documentation\/white_papers\/wp486-deep-learning-int8.pdf Last accessed Sept 12, 2021 . , Xilinx. 2017. Deep Learning with INT8 Optimization on Xilinx Devices. https:\/\/www.xilinx.com\/support\/documentation\/white_papers\/wp486-deep-learning-int8.pdf Last accessed Sept 12, 2021."},{"key":"e_1_3_2_2_48_1","volume-title":"Convolutional Neural Network with INT4 Optimization on Xilinx Devices. https:\/\/www.xilinx.com\/support\/documentation\/white_papers\/wp521--4bit-optimization.pdf Last accessed","year":"2021","unstructured":", Xilinx. 2020. Convolutional Neural Network with INT4 Optimization on Xilinx Devices. https:\/\/www.xilinx.com\/support\/documentation\/white_papers\/wp521--4bit-optimization.pdf Last accessed Sept 12, 2021 . , Xilinx. 2020. Convolutional Neural Network with INT4 Optimization on Xilinx Devices. https:\/\/www.xilinx.com\/support\/documentation\/white_papers\/wp521--4bit-optimization.pdf Last accessed Sept 12, 2021."},{"key":"e_1_3_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/3289602.3293902"},{"key":"e_1_3_2_2_50_1","volume-title":"PyHessian: Neural networks through the lens of the Hessian. arXiv preprint arXiv:1912.07145","author":"Yao Zhewei","year":"2019","unstructured":"019)]% yao2019pyhessian, Zhewei Yao , Amir Gholami , Kurt Keutzer , and Michael Mahoney . 2019. PyHessian: Neural networks through the lens of the Hessian. arXiv preprint arXiv:1912.07145 ( 2019 ). 019)]% yao2019pyhessian, Zhewei Yao, Amir Gholami, Kurt Keutzer, and Michael Mahoney. 2019. PyHessian: Neural networks through the lens of the Hessian. arXiv preprint arXiv:1912.07145 (2019)."},{"key":"e_1_3_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/2684746.2689060"},{"key":"e_1_3_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01237-3_23"},{"key":"e_1_3_2_2_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3240765.3240801"},{"key":"e_1_3_2_2_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/3020078.3021741"},{"key":"e_1_3_2_2_55_1","volume-title":"Incremental network quantization: Towards lossless cnns with low-precision weights. arXiv preprint arXiv:1702.03044","author":"Zhou Aojun","year":"2017","unstructured":"017)]% zhou2017incremental, Aojun Zhou , Anbang Yao , Yiwen Guo , Lin Xu , and Yurong Chen . 2017. Incremental network quantization: Towards lossless cnns with low-precision weights. arXiv preprint arXiv:1702.03044 ( 2017 ). 017)]% zhou2017incremental, Aojun Zhou, Anbang Yao, Yiwen Guo, Lin Xu, and Yurong Chen. 2017. Incremental network quantization: Towards lossless cnns with low-precision weights. arXiv preprint arXiv:1702.03044 (2017)."},{"key":"e_1_3_2_2_56_1","volume-title":"Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint arXiv:1606.06160","author":"Zhou Shuchang","year":"2016","unstructured":"016)]% zhou2016dorefa, Shuchang Zhou , Yuxin Wu , Zekun Ni , Xinyu Zhou , He Wen , and Yuheng Zou . 2016 . Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint arXiv:1606.06160 (2016). 016)]% zhou2016dorefa, Shuchang Zhou, Yuxin Wu, Zekun Ni, Xinyu Zhou, He Wen, and Yuheng Zou. 2016. Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint arXiv:1606.06160 (2016)."},{"key":"e_1_3_2_2_57_1","volume-title":"International Conference on Learning Representations (ICLR) .","author":"Zhu Chenzhuo","year":"2017","unstructured":"017)]% zhu2016trained, Chenzhuo Zhu , Song Han , Huizi Mao , and William J Dally . 2017 . Trained ternary quantization . In International Conference on Learning Representations (ICLR) . 017)]% zhu2016trained, Chenzhuo Zhu, Song Han, Huizi Mao, and William J Dally. 2017. Trained ternary quantization. In International Conference on Learning Representations (ICLR) ."}],"event":{"name":"FPGA '22: The 2022 ACM\/SIGDA International Symposium on Field-Programmable Gate Arrays","location":"Virtual Event USA","acronym":"FPGA '22","sponsor":["SIGDA ACM Special Interest Group on Design Automation"]},"container-title":["Proceedings of the 2022 ACM\/SIGDA International Symposium on Field-Programmable Gate Arrays"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3490422.3502364","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3490422.3502364","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3490422.3502364","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:31:03Z","timestamp":1750188663000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3490422.3502364"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,2,11]]},"references-count":57,"alternative-id":["10.1145\/3490422.3502364","10.1145\/3490422"],"URL":"https:\/\/doi.org\/10.1145\/3490422.3502364","relation":{},"subject":[],"published":{"date-parts":[[2022,2,11]]},"assertion":[{"value":"2022-02-11","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}