{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T20:09:44Z","timestamp":1762027784428,"version":"build-2065373602"},"reference-count":29,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,10,1]],"date-time":"2020-10-01T00:00:00Z","timestamp":1601510400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,10,1]],"date-time":"2020-10-01T00:00:00Z","timestamp":1601510400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,10,1]],"date-time":"2020-10-01T00:00:00Z","timestamp":1601510400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,10]]},"DOI":"10.1109\/sips50750.2020.9195219","type":"proceedings-article","created":{"date-parts":[[2020,9,23]],"date-time":"2020-09-23T20:55:00Z","timestamp":1600894500000},"page":"1-6","source":"Crossref","is-referenced-by-count":9,"title":["Knowledge Distillation for Optimization of Quantized Deep Neural Networks"],"prefix":"10.1109","author":[{"given":"Sungho","family":"Shin","sequence":"first","affiliation":[]},{"given":"Yoonho","family":"Boo","sequence":"additional","affiliation":[]},{"given":"Wonyong","family":"Sung","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"crossref","first-page":"535","DOI":"10.1145\/1150402.1150464","article-title":"Model compression","author":"bucilu?","year":"2006","journal-title":"Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472809"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3209996"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7953145"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2019.2903943"},{"article-title":"Fitnets: Hints for thin deep nets","year":"2014","author":"romero","key":"ref15"},{"article-title":"Knowledge distillation using unlabeled mismatched images","year":"2017","author":"kulkarni","key":"ref16"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00409"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.754"},{"article-title":"Apprentice: Using knowledge distillation techniques to improve low-precision network accuracy","year":"0","author":"mishra","key":"ref19"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/3336191.3371792"},{"key":"ref4","first-page":"1","article-title":"Quantized neural networks: Training neural networks with low precision weights and activations","volume":"18","author":"hubara","year":"2017","journal-title":"The Journal of Machine Learning Research"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8682462"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/MCI.2018.2840738"},{"key":"ref6","article-title":"Alternating multi-bit quantization for recurrent neural networks","author":"xu","year":"2018","journal-title":"International Conference on Learning Representations (ICLR)"},{"article-title":"Generative knowledge transfer for neural language models","year":"2016","author":"shin","key":"ref29"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/SiPS.2014.6986082"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1007\/s11390-017-1750-y"},{"article-title":"Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients","year":"2016","author":"zhou","key":"ref7"},{"key":"ref2","first-page":"173","article-title":"Deep speech 2: End-to-end speech recognition in english and mandarin","author":"amodei","year":"2016","journal-title":"International Conference on Machine Learning"},{"article-title":"Distilling the knowledge in a neural network","year":"2015","author":"hinton","key":"ref9"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"article-title":"Model compression via distillation and quantization","year":"0","author":"polino","key":"ref20"},{"article-title":"Improved knowledge distillation via teacher assistant: Bridging the gap between student and teacher","year":"2019","author":"mirzadeh","key":"ref22"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00826"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7952347"},{"key":"ref23","first-page":"525","article-title":"Xnor-net: Imagenet classification using binary convolutional neural networks","author":"rastegari","year":"2016","journal-title":"Proceedings of the European Conference on Computer Vision (ECCV)"},{"key":"ref26","doi-asserted-by":"crossref","DOI":"10.5244\/C.30.87","article-title":"Wide residual networks","author":"zagoruyko","year":"2016"},{"article-title":"Resiliency of deep neural networks under quantization","year":"2015","author":"sung","key":"ref25"}],"event":{"name":"2020 IEEE Workshop on Signal Processing Systems (SiPS)","start":{"date-parts":[[2020,10,20]]},"location":"Coimbra, Portugal","end":{"date-parts":[[2020,10,22]]}},"container-title":["2020 IEEE Workshop on Signal Processing Systems (SiPS)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9187068\/9195186\/09195219.pdf?arnumber=9195219","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,30]],"date-time":"2022-06-30T15:18:59Z","timestamp":1656602339000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9195219\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,10]]},"references-count":29,"URL":"https:\/\/doi.org\/10.1109\/sips50750.2020.9195219","relation":{},"subject":[],"published":{"date-parts":[[2020,10]]}}}