{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,30]],"date-time":"2025-09-30T11:08:31Z","timestamp":1759230511335},"reference-count":40,"publisher":"IEEE","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2018,7]]},"DOI":"10.1109\/ijcnn.2018.8489764","type":"proceedings-article","created":{"date-parts":[[2018,10,19]],"date-time":"2018-10-19T22:25:09Z","timestamp":1539987909000},"page":"1-8","source":"Crossref","is-referenced-by-count":20,"title":["PruNet: Class-Blind Pruning Method For Deep Neural Networks"],"prefix":"10.1109","author":[{"given":"Alberto","family":"Marchisio","sequence":"first","affiliation":[]},{"given":"Muhammad Abdullah","family":"Hanif","sequence":"additional","affiliation":[]},{"given":"Maurizio","family":"Martina","sequence":"additional","affiliation":[]},{"given":"Muhammad","family":"Shafique","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","article-title":"To prune, or not to prune: exploring the efficacy of pruning for model compression","author":"zhu","year":"2017","journal-title":"arXiv preprint arXiv 1710 01878"},{"key":"ref38","article-title":"Deep image: scaling up image recognition","author":"wu","year":"2015","journal-title":"arXiv preprint arXiv 1501 02876"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref32","article-title":"Dropout: A simple way to prevent neural networks from overtting","author":"srivastava","year":"2014","journal-title":"Journal of Machine Learning Research"},{"key":"ref31","article-title":"Very deep convolutional networks for large-scale image recognition","author":"simonyan","year":"2015","journal-title":"ICLRE"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ISVLSI.2017.124"},{"key":"ref37","article-title":"Learning structured sparsity in deep neural networks","author":"wen","year":"2016","journal-title":"NIPS"},{"key":"ref36","article-title":"Accelerating deep convolutional networks using low-precision and sparsity","author":"venkatesh","year":"2016","journal-title":"arXiv preprint arXiv 1610 01292"},{"key":"ref35","article-title":"Improving the speed of neural networks on CPUs","author":"vanhoucke","year":"2011","journal-title":"NIPS"},{"key":"ref34","article-title":"Soft Weight-Sharing for Neural Network Compression","author":"ullrich","year":"2017","journal-title":"ICLRE"},{"key":"ref10","article-title":"Dynamic network surgery for efficient dnns","author":"guo","year":"2016","journal-title":"NIPS"},{"year":"0","key":"ref40"},{"key":"ref11","article-title":"Deep learning with limited numerical precision","author":"gupta","year":"2015","journal-title":"CoRR"},{"key":"ref12","article-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding","author":"han","year":"2015","journal-title":"ICLRE"},{"key":"ref13","article-title":"Learning both weights and connections for efficient neural network","author":"han","year":"2015","journal-title":"NIPS"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1145\/3061639.3062306"},{"key":"ref15","article-title":"Error resilience analysis for systematically employing approximate computing in convolutional neural networks","author":"hanif","year":"2018","journal-title":"DATE"},{"key":"ref16","article-title":"Second order derivatives for network pruning: Optimal brain surgeon","author":"hassibi","year":"1993","journal-title":"NIPS"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.155"},{"key":"ref19","article-title":"Distilling the knowledge in a neural network","author":"hinton","year":"2015","journal-title":"NIPS"},{"key":"ref28","article-title":"Exploring sparsity in recurrent neural networks","author":"narang","year":"2017","journal-title":"CoRR"},{"key":"ref4","article-title":"Binarynet: Training deep neural networks with weights and activations constrained to +1 or 1","author":"courbariaux","year":"2016","journal-title":"arXiv preprint arXiv 1602 04875"},{"key":"ref27","article-title":"Variational dropout sparsifies deep neural networks","author":"molchanov","year":"2017","journal-title":"arXiv preprint arXiv 1701 05369"},{"key":"ref3","article-title":"Compressing neural networks with the hashing trick","author":"chen","year":"2015","journal-title":"ICML"},{"key":"ref6","article-title":"Training deep neural networks with low precision multiplications","author":"courbariaux","year":"2014","journal-title":"arXiv preprint arXiv 1412 7024"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/K16-1029"},{"key":"ref5","article-title":"Binaryconnect: Training deep neural networks with binary weights during propagations","author":"courbariaux","year":"2015","journal-title":"NIPS"},{"key":"ref8","article-title":"Learning to prune deep neural networks via layer-wise optimal brain surgeon","author":"dong","year":"2017","journal-title":"arXiv preprint arXiv 1705 07565"},{"key":"ref7","article-title":"Exploiting linear structure within convolutional networks for efficient evaluation","author":"denton","year":"2014","journal-title":"NIPS"},{"key":"ref2","article-title":"The power of sparsity in convolutional neural networks","author":"changpinyo","year":"2017","journal-title":"ICLRE"},{"key":"ref9","article-title":"Improved Bayesian Compression","author":"federici","year":"2017","journal-title":"NIPS"},{"key":"ref1","article-title":"Structured pruning of deep convolutional neural networks","author":"anwar","year":"2015","journal-title":"arXiv preprint arXiv 1512 08571"},{"key":"ref20","article-title":"Speeding up convolutional neural networks with low rank expansions","author":"jaderberg","year":"2014","journal-title":"NIPS"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"ref21","article-title":"Imagenet classification with deep convolutional neural networks","author":"krizhevsky","year":"2012","journal-title":"NIPS"},{"key":"ref24","article-title":"Neural networks with few multiplications","author":"lin","year":"2016","journal-title":"ICLRE"},{"key":"ref23","article-title":"Pruning filters for efficient convnets","author":"li","year":"2016","journal-title":"arXiv preprint arXiv 1608 08710"},{"key":"ref26","article-title":"Bayesian compression for deep learning","author":"louizos","year":"2017","journal-title":"NIPS"},{"key":"ref25","article-title":"Fixed point quantization of deep convolutional networks","author":"lin","year":"2015","journal-title":"arXiv preprint arXiv 1511 05271"}],"event":{"name":"2018 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2018,7,8]]},"location":"Rio de Janeiro","end":{"date-parts":[[2018,7,13]]}},"container-title":["2018 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8465565\/8488986\/08489764.pdf?arnumber=8489764","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2020,8,24]],"date-time":"2020-08-24T04:57:51Z","timestamp":1598245071000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8489764\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,7]]},"references-count":40,"URL":"https:\/\/doi.org\/10.1109\/ijcnn.2018.8489764","relation":{},"subject":[],"published":{"date-parts":[[2018,7]]}}}