{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T16:55:09Z","timestamp":1773248109628,"version":"3.50.1"},"reference-count":27,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,7,9]],"date-time":"2023-07-09T00:00:00Z","timestamp":1688860800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,7,9]],"date-time":"2023-07-09T00:00:00Z","timestamp":1688860800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,7,9]]},"DOI":"10.1109\/dac56929.2023.10247982","type":"proceedings-article","created":{"date-parts":[[2023,9,15]],"date-time":"2023-09-15T13:31:31Z","timestamp":1694784691000},"page":"1-6","source":"Crossref","is-referenced-by-count":6,"title":["CSQ: Growing Mixed-Precision Quantization Scheme with Bi-level Continuous Sparsification"],"prefix":"10.1109","author":[{"given":"Lirui","family":"Xiao","sequence":"first","affiliation":[{"name":"Nanjing University"}]},{"given":"Huanrui","family":"Yang","sequence":"additional","affiliation":[{"name":"University of California,Berkeley"}]},{"given":"Zhen","family":"Dong","sequence":"additional","affiliation":[{"name":"University of California,Berkeley"}]},{"given":"Kurt","family":"Keutzer","sequence":"additional","affiliation":[{"name":"University of California,Berkeley"}]},{"given":"Li","family":"Du","sequence":"additional","affiliation":[{"name":"Nanjing University"}]},{"given":"Shanghang","family":"Zhang","sequence":"additional","affiliation":[{"name":"Peking University"}]}],"member":"263","reference":[{"key":"ref13","article-title":"Growing efficient deep networks by structured continuous sparsification","author":"yuan","year":"2020"},{"key":"ref12","first-page":"11380","article-title":"Winning the lottery with continuous sparsification","volume":"33","author":"savarese","year":"2020","journal-title":"NeurIPS"},{"key":"ref15","article-title":"Ternary weight networks","author":"li","year":"2016"},{"key":"ref14","first-page":"525","article-title":"Xnor-net: Imagenet classification using binary convolutional neural networks","author":"rastegari","year":"2016","journal-title":"ECCV"},{"key":"ref11","article-title":"Estimating or propagating gradients through stochastic neurons for conditional computation","author":"bengio","year":"2013"},{"key":"ref10","first-page":"18518","article-title":"Hawq-v2: Hessian aware trace-weighted quantization of neural networks","volume":"33","author":"dong","year":"2020","journal-title":"NeurIPS"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00474"},{"key":"ref17","first-page":"3288","article-title":"Bayesian compression for deep learning","author":"louizos","year":"2017","journal-title":"NeurIPS"},{"key":"ref16","article-title":"Learning sparse neural networks through l_0 regularization","author":"louizos","year":"2017"},{"key":"ref19","article-title":"Very deep convolutional networks for large-scale image recognition","author":"simonyan","year":"2014"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01318"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58520-4_23"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1145\/3370748.3406585"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00156"},{"key":"ref20","article-title":"Learning multiple layers of features from tiny images","year":"2009","journal-title":"Technical Report"},{"key":"ref22","first-page":"11875","article-title":"Hawq-v3: Dyadic neural network quantization","author":"yao","year":"2021","journal-title":"ICML"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref27","article-title":"Model compression via distillation and quantization","author":"polino","year":"2018"},{"key":"ref8","article-title":"Bsq: Exploring bit-level sparsity for mixed-precision neural network quantization","author":"yang","year":"2021"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00038"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00881"},{"key":"ref4","article-title":"Pact: Parameterized clipping activation for quantized neural networks","author":"choi","year":"2018"},{"key":"ref3","article-title":"Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients","author":"zhou","year":"2016"},{"key":"ref6","article-title":"1.1 computing&#x2019;s energy problem (and what we can do about it)","author":"horowitz","year":"2014","journal-title":"ISSCC"},{"key":"ref5","first-page":"365","article-title":"Lq-nets: Learned quantization for highly accurate and compact deep neural networks","author":"zhang","year":"2018","journal-title":"ECCV"}],"event":{"name":"2023 60th ACM\/IEEE Design Automation Conference (DAC)","location":"San Francisco, CA, USA","start":{"date-parts":[[2023,7,9]]},"end":{"date-parts":[[2023,7,13]]}},"container-title":["2023 60th ACM\/IEEE Design Automation Conference (DAC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10247654\/10247655\/10247982.pdf?arnumber=10247982","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,10,2]],"date-time":"2023-10-02T13:41:37Z","timestamp":1696254097000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10247982\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,9]]},"references-count":27,"URL":"https:\/\/doi.org\/10.1109\/dac56929.2023.10247982","relation":{},"subject":[],"published":{"date-parts":[[2023,7,9]]}}}