{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,22]],"date-time":"2026-03-22T22:42:41Z","timestamp":1774219361674,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":24,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,7,10]],"date-time":"2022-07-10T00:00:00Z","timestamp":1657411200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2018YFB1403400"],"award-info":[{"award-number":["2018YFB1403400"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61834006 and 62102257"],"award-info":[{"award-number":["61834006 and 62102257"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,7,10]]},"DOI":"10.1145\/3489517.3530660","type":"proceedings-article","created":{"date-parts":[[2022,8,23]],"date-time":"2022-08-23T23:19:29Z","timestamp":1661296769000},"page":"259-264","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":13,"title":["EBSP"],"prefix":"10.1145","author":[{"given":"Fangxin","family":"Liu","sequence":"first","affiliation":[{"name":"Shanghai Jiao Tong University and Shanghai Qi Zhi Institute"}]},{"given":"Wenbo","family":"Zhao","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University"}]},{"given":"Zongwu","family":"Wang","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University"}]},{"given":"Yongbiao","family":"Chen","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University"}]},{"given":"Zhezhi","family":"He","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University"}]},{"given":"Naifeng","family":"Jing","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University"}]},{"given":"Xiaoyao","family":"Liang","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University"}]},{"given":"Li","family":"Jiang","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University"}]}],"member":"320","published-online":{"date-parts":[[2022,8,23]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"crossref","unstructured":"Vahideh Akhlaghi et al. 2018. SnaPEA: Predictive Early Activation for Reducing Computation in Deep Convolutional Neural Networks. In ISCA.","DOI":"10.1109\/ISCA.2018.00061"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"Rajeev Balasubramonian et al. 2017. CACTI 7: New Tools for Interconnect Exploration in Innovative Off-Chip Memories. TACO (2017).","DOI":"10.1145\/3085572"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"Sung-En Chang et al. 2021. Mix and Match: A novel FPGA-centric deep neural network quantization framework. In HPCA. IEEE.","DOI":"10.1109\/HPCA51647.2021.00027"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3007787.3001177"},{"key":"e_1_3_2_1_5_1","unstructured":"Synopsys Design Compiler. 2019. [Online]. Available: https:\/\/www.synopsys.com\/support\/training\/rtlsynthesis\/design-compiler-rtl-synthesis.html."},{"key":"e_1_3_2_1_6_1","volume-title":"Imagenet: A large-scale hierarchical image database. In CVPR. Ieee, 248--255.","author":"Jia Deng","year":"2009","unstructured":"Jia Deng et al. 2009. Imagenet: A large-scale hierarchical image database. In CVPR. Ieee, 248--255."},{"key":"e_1_3_2_1_7_1","first-page":"4","article-title":"2020. Model compression and hardware acceleration for neural networks: A comprehensive survey","volume":"108","author":"Lei Deng","year":"2020","unstructured":"Lei Deng et al. 2020. Model compression and hardware acceleration for neural networks: A comprehensive survey. Proc. IEEE 108, 4 (2020), 485--532.","journal-title":"Proc. IEEE"},{"key":"e_1_3_2_1_8_1","unstructured":"Steven K Esser et al. 2020. Learned step size quantization. In ICLR."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"crossref","unstructured":"Michael Gautschi Michael Schaffner et al. 2016. 4.6 A 65nm CMOS 6.4-to-29.2 pJ\/FLOP@ 0.8 V shared logarithmic floating point unit for acceleration of nonlinear function kernels in a tightly coupled processor cluster. In ISSCC. IEEE.","DOI":"10.1109\/ISSCC.2016.7417917"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"crossref","unstructured":"Sumanth Gudaparthi et al. 2019. Wire-Aware Architecture and Dataflow for CNN Accelerators. In MICRO.","DOI":"10.1145\/3352460.3358316"},{"key":"e_1_3_2_1_11_1","volume-title":"Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding. In ICLR.","author":"Song Han","year":"2016","unstructured":"Song Han et al. 2016. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding. In ICLR."},{"key":"e_1_3_2_1_12_1","volume-title":"EIE: Efficient inference engine on compressed deep neural network. ACM SIGARCH Computer Architecture News","author":"Song Han","year":"2016","unstructured":"Song Han et al. 2016. EIE: Efficient inference engine on compressed deep neural network. ACM SIGARCH Computer Architecture News (2016)."},{"key":"e_1_3_2_1_13_1","unstructured":"Kaiming He et al. 2016. Deep residual learning for image recognition. In CVPR."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"Benoit Jacob etal. 2018. Quantization and training of neural networks for efficient integer-arithmetic-only inference. In CVPR.","DOI":"10.1109\/CVPR.2018.00286"},{"key":"e_1_3_2_1_15_1","unstructured":"Alex K. et al. 2009. Learning multiple layers of features from tiny images."},{"key":"e_1_3_2_1_16_1","volume-title":"Hinton","author":"Krizhevsky Alex","year":"2017","unstructured":"Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. 2017. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM (2017)."},{"key":"e_1_3_2_1_17_1","unstructured":"Fangxin Liu et al. 2021. Improving Neural Network Efficiency via Post-training Quantization with Adaptive Floating-Point. In ICCV."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"crossref","unstructured":"S\u00e9bastien Marcel and Yann Rodriguez. 2010. Torchvision the Machine-Vision Package of Torch. In MM.","DOI":"10.1145\/1873951.1874254"},{"key":"e_1_3_2_1_19_1","volume-title":"Patdnn: Achieving real-time DNN execution on mobile devices with pattern-based weight pruning. In ASPLOS.","author":"Wei Niu","year":"2020","unstructured":"Wei Niu et al. 2020. Patdnn: Achieving real-time DNN execution on mobile devices with pattern-based weight pruning. In ASPLOS."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"crossref","unstructured":"Eunhyeok Park et al. 2018. Energy-efficient neural network accelerator based on outlier-aware low-precision computation. In ISCA.","DOI":"10.1109\/ISCA.2018.00063"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"crossref","unstructured":"Akshay Krishna Ramanathan et al. 2020. Look-up table based energy efficient processing in cache support for neural network acceleration. In MICRO. IEEE.","DOI":"10.1109\/MICRO50266.2020.00020"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"crossref","unstructured":"Mark Sandler Andrew Howard Menglong Zhu et al. 2018. Mobilenetv2: Inverted residuals and linear bottlenecks. In CVPR. 4510--4520.","DOI":"10.1109\/CVPR.2018.00474"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"crossref","unstructured":"Sayeh Sharify et al. 2019. Laconic Deep Learning Inference Acceleration. In ISCA (Phoenix Arizona) (ISCA '19). 304--317.","DOI":"10.1145\/3307650.3322255"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"crossref","unstructured":"Hardik Sharma et al. 2018. Bit fusion: Bit-level dynamically composable architecture for accelerating deep neural network. In ISCA.","DOI":"10.1109\/ISCA.2018.00069"}],"event":{"name":"DAC '22: 59th ACM\/IEEE Design Automation Conference","location":"San Francisco California","acronym":"DAC '22","sponsor":["SIGDA ACM Special Interest Group on Design Automation","IEEE CEDA"]},"container-title":["Proceedings of the 59th ACM\/IEEE Design Automation Conference"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3489517.3530660","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3489517.3530660","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:02:23Z","timestamp":1750186943000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3489517.3530660"}},"subtitle":["evolving bit sparsity patterns for hardware-friendly inference of quantized deep neural networks"],"short-title":[],"issued":{"date-parts":[[2022,7,10]]},"references-count":24,"alternative-id":["10.1145\/3489517.3530660","10.1145\/3489517"],"URL":"https:\/\/doi.org\/10.1145\/3489517.3530660","relation":{},"subject":[],"published":{"date-parts":[[2022,7,10]]},"assertion":[{"value":"2022-08-23","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}