{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,9]],"date-time":"2026-01-09T20:30:59Z","timestamp":1767990659875,"version":"3.49.0"},"reference-count":57,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,3,1]],"date-time":"2020-03-01T00:00:00Z","timestamp":1583020800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,3,1]],"date-time":"2020-03-01T00:00:00Z","timestamp":1583020800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,3,1]],"date-time":"2020-03-01T00:00:00Z","timestamp":1583020800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,3]]},"DOI":"10.1109\/wacv45572.2020.9093546","type":"proceedings-article","created":{"date-parts":[[2020,5,15]],"date-time":"2020-05-15T03:41:09Z","timestamp":1589514069000},"page":"3129-3138","source":"Crossref","is-referenced-by-count":17,"title":["Filter Distillation for Network Compression"],"prefix":"10.1109","author":[{"given":"Xavier","family":"Suau","sequence":"first","affiliation":[]},{"given":"uca","family":"Zappella","sequence":"additional","affiliation":[]},{"given":"Nicholas","family":"Apostoloff","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01237-3_19"},{"key":"ref38","first-page":"240","article-title":"Notes on regression and inheritance in the case of two parents","author":"pearson","year":"1895","journal-title":"Royal Society of London"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.541"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref31","article-title":"On the computational efficiency of training neural networks","author":"livni","year":"2014","journal-title":"NIPS"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.298"},{"key":"ref37","first-page":"1473","author":"nakkiran","year":"2015","journal-title":"Compressing deep neural networks using a rank-constrained topology"},{"key":"ref36","article-title":"Pruning convolutional neural networks for resource efficient inference","author":"molchanov","year":"2017","journal-title":"ICLRE"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.460"},{"key":"ref34","article-title":"Democratizing production-scale distributed deep learning","author":"ma","year":"2018","journal-title":"CoRR abs\/1810 00143"},{"key":"ref28","first-page":"598","article-title":"Optimal brain damage","author":"lecun","year":"1990","journal-title":"NIPS"},{"key":"ref27","article-title":"Speeding-up convolutional neural networks using fine-tuned cp-decomposition","author":"lebedev","year":"2014","journal-title":"ICLRE"},{"key":"ref29","article-title":"Pruning filters for efficient convnets","author":"li","year":"2017","journal-title":"ICLRE"},{"key":"ref2","volume":"abs 1711 2638","author":"alvarez","year":"2017","journal-title":"Compression-aware Training of Deep Networks"},{"key":"ref1","first-page":"3180","article-title":"Nettrim: Convex pruning of deep neural networks with performance guarantee","author":"aghasi","year":"2017","journal-title":"NIPS"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01246-5_46"},{"key":"ref22","first-page":"1","article-title":"Distilling the knowledge in a neural network","author":"hinton","year":"2014","journal-title":"NIPS"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.155"},{"key":"ref24","article-title":"Speeding up convolutional neural networks with low rank expansions","author":"jaderberg","year":"2014","journal-title":"BMVC"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1037\/h0071325"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1214\/aoms\/1177729694"},{"key":"ref25","article-title":"Learning multiple layers of features from tiny images","author":"krizhevsky","year":"2009","journal-title":"Master&#x2019;s thesis"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.78"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.521"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01237-3_12"},{"key":"ref56","author":"zagoruyko","year":"0","journal-title":"92 45% on cifar-10 in torch"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.15"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00958"},{"key":"ref53","first-page":"6359","article-title":"Restructuring of deep neural network acoustic models with singular value decomposition","author":"xue","year":"2013","journal-title":"ICASSP"},{"key":"ref52","article-title":"Deep k-means: Re-training and parameter sharing with harder cluster assignments for compressing deep convolutions","author":"wu","year":"2018","journal-title":"ICML"},{"key":"ref10","article-title":"DNN or k-nn: That is the generalize vs. memorize question","author":"cohen","year":"2018","journal-title":"CoRR"},{"key":"ref11","article-title":"Compressing neural networks using the variational information bottleneck","author":"dai","year":"2018","journal-title":"ICML"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472823"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref13","first-page":"2148","article-title":"Predicting parameters in deep learning","author":"denil","year":"2013","journal-title":"NIPS"},{"key":"ref14","first-page":"1269","article-title":"Exploiting linear structure within convolutional networks for efficient evaluation","author":"denton","year":"2014","journal-title":"NIPS"},{"key":"ref15","year":"2015","journal-title":"TensorFlow Large-Scale Machine Learning on Heterogeneous Systems"},{"key":"ref16","article-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding","author":"han","year":"2016","journal-title":"ICLRE"},{"key":"ref17","article-title":"Learning both weights and connections for efficient neural networks","author":"han","year":"2016","journal-title":"NIPS"},{"key":"ref18","first-page":"263","article-title":"Optimal brain surgeon: Extensions and performance comparisons","author":"hassibi","year":"1993","journal-title":"NIPS"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref4","article-title":"Do deep nets really need to be deep&#x0192;","author":"ba","year":"2014","journal-title":"NIPS"},{"key":"ref3","article-title":"On the optimization of deep networks: Implicit acceleration by overparam-eterization","author":"arora","year":"2018","journal-title":"NIPS"},{"key":"ref6","article-title":"learning-compression","author":"carreira-perpinan","year":"2018","journal-title":"algorithms for neural net pruning In CVPR"},{"key":"ref5","first-page":"535","article-title":"Model compression","author":"bucilua","year":"2006","journal-title":"SIGKDD"},{"key":"ref8","article-title":"Convolutional neural networks with low-rank regularization","author":"cheng","year":"2016","journal-title":"ICLRE"},{"key":"ref7","article-title":"Net2net: Accelerating learning via knowledge transfer","volume":"11","author":"chen","year":"2016","journal-title":"ICLRE"},{"key":"ref49","first-page":"2074","article-title":"Learning structured sparsity in deep neural networks","author":"wen","year":"2016","journal-title":"NIPS"},{"key":"ref9","article-title":"Reducing overfitting in deep networks by decorrelating representations","author":"cogswell","year":"2016","journal-title":"ICLRE"},{"key":"ref46","article-title":"Very deep convolutional networks for large-scale image recognition","author":"simonyan","year":"2015","journal-title":"ICLRE"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.1998.710701"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00821"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.5244\/C.29.31"},{"key":"ref42","article-title":"Xnor-net: Imagenet classification using binary convolutional neural networks","author":"rastegari","year":"2016","journal-title":"ECCV"},{"key":"ref41","article-title":"The loss surface of deep and wide neural networks","author":"quynh","year":"2017","journal-title":"ICML"},{"key":"ref44","article-title":"Fitnets: Hints for thin deep nets","author":"romero","year":"2015","journal-title":"ICLRE"},{"key":"ref43","article-title":"Regularizing cnns with locally constrained decorrelations","author":"rodr\u00edguez","year":"2017","journal-title":"ICLRE"}],"event":{"name":"2020 IEEE Winter Conference on Applications of Computer Vision (WACV)","location":"Snowmass Village, CO, USA","start":{"date-parts":[[2020,3,1]]},"end":{"date-parts":[[2020,3,5]]}},"container-title":["2020 IEEE Winter Conference on Applications of Computer Vision (WACV)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9087828\/9093261\/09093546.pdf?arnumber=9093546","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,30]],"date-time":"2022-06-30T15:18:14Z","timestamp":1656602294000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9093546\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,3]]},"references-count":57,"URL":"https:\/\/doi.org\/10.1109\/wacv45572.2020.9093546","relation":{},"subject":[],"published":{"date-parts":[[2020,3]]}}}