{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T08:36:37Z","timestamp":1768898197991,"version":"3.49.0"},"reference-count":76,"publisher":"MDPI AG","issue":"4","license":[{"start":{"date-parts":[[2021,11,17]],"date-time":"2021-11-17T00:00:00Z","timestamp":1637107200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Informatics"],"abstract":"<jats:p>Deep networks often possess a vast number of parameters, and their significant redundancy in parameterization has become a widely-recognized property. This presents significant challenges and restricts many deep learning applications, making the focus on reducing the complexity of models while maintaining their powerful performance. In this paper, we present an overview of popular methods and review recent works on compressing and accelerating deep neural networks. We consider not only pruning methods but also quantization methods, and low-rank factorization methods. This review also intends to clarify these major concepts, and highlights their characteristics, advantages, and shortcomings.<\/jats:p>","DOI":"10.3390\/informatics8040077","type":"journal-article","created":{"date-parts":[[2021,11,17]],"date-time":"2021-11-17T21:32:07Z","timestamp":1637184727000},"page":"77","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":32,"title":["Literature Review of Deep Network Compression"],"prefix":"10.3390","volume":"8","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1052-2657","authenticated-orcid":false,"given":"Ali","family":"Alqahtani","sequence":"first","affiliation":[{"name":"Department of Computer Science, Swansea University, Swansea SA2 8PP, UK"},{"name":"Department of Computer Science, King Khalid University, Abha 62529, Saudi Arabia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2701-8660","authenticated-orcid":false,"given":"Xianghua","family":"Xie","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Swansea University, Swansea SA2 8PP, UK"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8991-1190","authenticated-orcid":false,"given":"Mark W.","family":"Jones","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Swansea University, Swansea SA2 8PP, UK"}]}],"member":"1968","published-online":{"date-parts":[[2021,11,17]]},"reference":[{"key":"ref_1","unstructured":"Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep Learning, MIT Press."},{"key":"ref_2","unstructured":"Denton, E.L., Zaremba, W., Bruna, J., LeCun, Y., and Fergus, R. (2014, January 8\u201313). Exploiting linear structure within convolutional networks for efficient evaluation. Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada."},{"key":"ref_3","unstructured":"Simonyan, K., and Zisserman, A. (2015, January 7\u20139). Very deep convolutional networks for large-scale image recognition. Proceedings of the International Conference on Learning Representations, San Diego, CA, USA."},{"key":"ref_4","unstructured":"Molchanov, P., Tyree, S., Karras, T., Aila, T., and Kautz, J. (2017, January 24\u201326). Pruning convolutional neural networks for resource efficient inference. Proceedings of the International Conference on Learning Representations, Toulon, France."},{"key":"ref_5","unstructured":"Krizhevsky, A., Sutskever, I., and Hinton, G.E. (2012, January 3\u20136). Imagenet classification with deep convolutional neural networks. Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., and Rabinovich, A. (2015, January 7\u201312). Going deeper with convolutions. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Wu, J., Leng, C., Wang, Y., Hu, Q., and Cheng, J. (2016, January 27\u201330). Quantized convolutional neural networks for mobile devices. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.521"},{"key":"ref_9","unstructured":"Denil, M., Shakibi, B., Dinh, L., Ranzato, M., and De Freitas, N. (2013, January 5\u201310). Predicting parameters in deep learning. Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"2525","DOI":"10.1109\/TPAMI.2018.2858232","article-title":"ThiNet: Pruning CNN Filters for a Thinner Net","volume":"41","author":"Luo","year":"2019","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_11","unstructured":"Mozer, M.C., and Smolensky, P. (,  1988). Skeletonization: A technique for trimming the fat from a network via relevance assessment. Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"740","DOI":"10.1109\/72.248452","article-title":"Pruning algorithms: A survey","volume":"4","author":"Reed","year":"1993","journal-title":"IEEE Trans. Neural Netw."},{"key":"ref_13","unstructured":"LeCun, Y., Denker, J.S., and Solla, S.A. (1990, January 26\u201329). Optimal brain damage. Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA."},{"key":"ref_14","unstructured":"Hassibi, B., and Stork, D.G. (,  1993). Second order derivatives for network pruning: Optimal brain surgeon. Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Weigend, A.S., Rumelhart, D.E., and Huberman, B.A. (1991, January 8\u201312). Generalization by weight-elimination applied to currency exchange rate prediction. Proceedings of the IEEE International Joint Conference on Neural Networks, Seattle, WA, USA.","DOI":"10.1109\/IJCNN.1991.170743"},{"key":"ref_16","unstructured":"Hanson, S., and Pratt, L. (,  1988). Comparing biases for minimal network construction with back-propagation. Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Weigend, A.S., Rumelhart, D.E., and Huberman, B.A. (1991). Back-propagation, weight-elimination and time series prediction. Connectionist Models, Morgan Kaufmann.","DOI":"10.1016\/B978-1-4832-1448-1.50016-0"},{"key":"ref_18","unstructured":"Arora, S., Ge, R., Neyshabur, B., and Zhang, Y. (2018, January 10\u201315). Stronger generalization bounds for deep nets via a compression approach. Proceedings of the International Conference on Machine Learning, Stockholm, Sweden."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Li, Z., Zhang, Z., Zhao, H., Wang, R., Chen, K., Utiyama, M., and Sumita, E. (2021). Text Compression-aided Transformer Encoding. IEEE Trans. Pattern Anal. Mach. Intell., 1.","DOI":"10.1109\/TPAMI.2021.3058341"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Amich, M., Luca, P.D., and Fiscale, S. (2020, January 5\u20138). Accelerated implementation of FQSqueezer novel genomic compression method. Proceedings of the International Symposium on Parallel and Distributed Computing, Warsaw, Poland.","DOI":"10.1109\/ISPDC51135.2020.00030"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"1309","DOI":"10.1109\/83.855427","article-title":"The LOCO-I lossless image compression algorithm: Principles and standardization into JPEG-LS","volume":"9","author":"Weinberger","year":"2000","journal-title":"IEEE Trans. Image Process."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Nagoor, O., Whittle, J., Deng, J., Mora, B., and Jones, M.W. (2021, January 10\u201315). MedZip: 3D Medical Images Lossless Compressor Using Recurrent Neural Network (LSTM). Proceedings of the International Conference on Pattern Recognition, Milan, Italy.","DOI":"10.1109\/ICPR48806.2021.9413341"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Nagoor, O., Whittle, J., Deng, J., Mora, B., and Jones, M.W. (2020, January 25\u201328). Lossless Compression For Volumetric Medical Images Using Deep Neural Network With Local Sampling. Proceedings of the IEEE International Conference on Image Processing, Abu Dhabi, United Arab Emirates.","DOI":"10.1109\/ICIP40778.2020.9191031"},{"key":"ref_24","unstructured":"Han, S., Pool, J., Tran, J., and Dally, W. (2015, January 7\u201312). Learning both weights and connections for efficient neural network. Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada."},{"key":"ref_25","unstructured":"Liu, Z., Sun, M., Zhou, T., Huang, G., and Darrell, T. (2019, January 6\u20139). Rethinking the value of network pruning. Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"2383","DOI":"10.1038\/s41467-018-04316-3","article-title":"Scalable training of artificial neural networks with adaptive sparse connectivity inspired by network science","volume":"9","author":"Mocanu","year":"2018","journal-title":"Nat. Commun."},{"key":"ref_27","unstructured":"Wen, W., Wu, C., Wang, Y., Chen, Y., and Li, H. (2016, January 5\u201310). Learning structured sparsity in deep neural networks. Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain."},{"key":"ref_28","unstructured":"Frankle, J., and Carbin, M. (2019, January 6\u20139). The lottery ticket hypothesis: Finding sparse, trainable neural networks. Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA."},{"key":"ref_29","unstructured":"Frankle, J., Dziugaite, G.K., Roy, D.M., and Carbin, M. (2019). Stabilizing the lottery ticket hypothesis. arXiv."},{"key":"ref_30","unstructured":"Morcos, A., Yu, H., Paganini, M., and Tian, Y. (2019, January 10\u201312). One ticket to win them all: Generalizing lottery ticket initializations across datasets and optimizers. Proceedings of the Neural Information Processing Systems, Vancouver, BC, Canada."},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Hubens, N., Mancas, M., Decombas, M., Preda, M., Zaharia, T., Gosselin, B., and Dutoit, T. (2020, January 7\u201312). An Experimental Study of the Impact of Pre-Training on the Pruning of a Convolutional Neural Network. Proceedings of the International Conference on Applications of Intelligent Systems, Las Palmas de Gran Canaria, Spain.","DOI":"10.1145\/3378184.3378224"},{"key":"ref_32","unstructured":"Zhou, H., Lan, J., Liu, R., and Yosinski, J. (2019, January 10\u201312). Deconstructing lottery tickets: Zeros, signs, and the supermask. Proceedings of the Neural Information Processing Systems, Vancouver, BC, Canada."},{"key":"ref_33","unstructured":"Yu, H., Edunov, S., Tian, Y., and Morcos, A.S. (2020). Playing the lottery with rewards and multiple languages: Lottery tickets in RL and NLP. arXiv."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Lebedev, V., and Lempitsky, V. (2016, January 27\u201330). Fast ConvNets using group-wise brain damage. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.280"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Zhou, H., Alvarez, J.M., and Porikli, F. (2016, January 11\u201314). Less is more: Towards compact CNNs. Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46493-0_40"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"He, T., Fan, Y., Qian, Y., Tan, T., and Yu, K. (2014, January 4\u20139). Reshaping deep neural network for fast decoding by node-pruning. Proceedings of the International Conference on Acoustics, Speech and Signal Processing, Florence, Italy.","DOI":"10.1109\/ICASSP.2014.6853595"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Alqahtani, A., Xie, X., Essa, E., and Jones, M.W. (2021, January 10\u201315). Neuron-based Network Pruning Based on Majority Voting. Proceedings of the International Conference on Pattern Recognition, Milan, Italy.","DOI":"10.1109\/ICPR48806.2021.9412897"},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Srinivas, S., and Babu, R.V. (2015, January 7\u201310). Data-free Parameter Pruning for Deep Neural Networks. Proceedings of the British Machine Vision Conference, Swansea, UK.","DOI":"10.5244\/C.29.31"},{"key":"ref_39","unstructured":"Mariet, Z., and Sra, S. (2016, January 2\u20134). Diversity networks: Neural network compression using determinantal point processes. Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico."},{"key":"ref_40","unstructured":"Li, H., Kadav, A., Durdanovic, I., Samet, H., and Graf, H.P. (2017, January 24\u201326). Pruning filters for efficient ConvNets. Proceedings of the International Conference on Learning Representations, Toulon, France."},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1016\/j.sigpro.2018.10.019","article-title":"Channel pruning based on mean gradient for accelerating Convolutional Neural Networks","volume":"156","author":"Liu","year":"2019","journal-title":"Signal Process."},{"key":"ref_42","doi-asserted-by":"crossref","first-page":"2163","DOI":"10.1109\/ACCESS.2015.2494536","article-title":"Channel-level acceleration of deep face representations","volume":"3","author":"Polyak","year":"2015","journal-title":"IEEE Access"},{"key":"ref_43","unstructured":"Luo, J.H., and Wu, J. (2017). An entropy-based pruning method for cnn compression. arXiv."},{"key":"ref_44","unstructured":"Hu, H., Peng, R., Tai, Y.W., and Tang, C.K. (2016). Network trimming: A data-driven neuron pruning approach towards efficient deep architectures. arXiv."},{"key":"ref_45","unstructured":"Liu, L., Zhang, S., Kuang, Z., Zhou, A., Xue, J., Wang, X., Chen, Y., Yang, W., Liao, Q., and Zhang, W. (2021, January 18\u201324). Group Fisher Pruning for Practical Network Compression. Proceedings of the International Conference on Machine Learning, Virtual, Vienna, Austria."},{"key":"ref_46","doi-asserted-by":"crossref","first-page":"107899","DOI":"10.1016\/j.patcog.2021.107899","article-title":"Pruning by Explaining: A Novel Criterion for Deep Neural Network Pruning","volume":"115","author":"Yeom","year":"2021","journal-title":"Pattern Recognit."},{"key":"ref_47","doi-asserted-by":"crossref","unstructured":"Bach, S., Binder, A., Montavon, G., Klauschen, F., Muller, K.R., and Samek, W. (2015). On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS ONE, 10.","DOI":"10.1371\/journal.pone.0130140"},{"key":"ref_48","doi-asserted-by":"crossref","first-page":"103220","DOI":"10.1016\/j.cviu.2021.103220","article-title":"Pruning CNN filters via quantifying the importance of deep visual representations","volume":"208","author":"Alqahtani","year":"2021","journal-title":"Comput. Vis. Image Underst."},{"key":"ref_49","doi-asserted-by":"crossref","unstructured":"Ding, X., Ding, G., Guo, Y., and Han, J. (2019, January 15\u201320). Centripetal SGD for pruning very deep convolutional networks with complicated structure. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00508"},{"key":"ref_50","doi-asserted-by":"crossref","unstructured":"He, Y., Kang, G., Dong, X., Fu, Y., and Yang, Y. (2018, January 13\u201319). Soft filter pruning for accelerating deep convolutional neural networks. Proceedings of the International Joint Conference on Artificial Intelligence, Stockholm, Sweden.","DOI":"10.24963\/ijcai.2018\/309"},{"key":"ref_51","doi-asserted-by":"crossref","unstructured":"He, Y., Liu, P., Wang, Z., Hu, Z., and Yang, Y. (2019, January 15\u201320). Filter pruning via geometric median for deep convolutional neural networks acceleration. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00447"},{"key":"ref_52","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mu, H., Zhang, X., Guo, Z., Yang, X., Cheng, K.T., and Sun, J. (2019, January 27\u201328). Metapruning: Meta learning for automatic neural network channel pruning. Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea.","DOI":"10.1109\/ICCV.2019.00339"},{"key":"ref_53","unstructured":"You, Z., Yan, K., Ye, J., Ma, M., and Wang, P. (2019, January 10\u201312). Gate decorator: Global filter pruning method for accelerating deep convolutional neural networks. Proceedings of the Neural Information Processing Systems, Vancouver, BC, Canada."},{"key":"ref_54","doi-asserted-by":"crossref","first-page":"107461","DOI":"10.1016\/j.patcog.2020.107461","article-title":"Autopruner: An end-to-end trainable filter pruning method for efficient deep model inference","volume":"107","author":"Luo","year":"2020","journal-title":"Pattern Recognit."},{"key":"ref_55","doi-asserted-by":"crossref","unstructured":"Rastegari, M., Ordonez, V., Redmon, J., and Farhadi, A. (2016, January 11\u201314). XNOR-Net: Imagenet classification using binary convolutional neural networks. Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46493-0_32"},{"key":"ref_56","unstructured":"Zhao, Y., Gao, X., Bates, D., Mullins, R., and Xu, C.Z. (2019, January 8\u201314). Focused quantization for sparse CNNs. Proceedings of the Neural Information Processing Systems, Vancouver, BC, Canada."},{"key":"ref_57","unstructured":"Zhou, A., Yao, A., Guo, Y., Xu, L., and Chen, Y. (2017). Incremental network quantization: Towards lossless CNNs with low-precision weights. arXiv."},{"key":"ref_58","unstructured":"Gong, Y., Liu, L., Yang, M., and Bourdev, L. (2014). Compressing deep convolutional networks using vector quantization. arXiv."},{"key":"ref_59","unstructured":"Vanhoucke, V., Senior, A., and Mao, M.Z. (2011, January 16). Improving the speed of neural networks on CPUs. Proceedings of the NIPS Workshop on Deep Learning and Unsupervised Feature Learning, Grenada, Spain."},{"key":"ref_60","unstructured":"Gupta, S., Agrawal, A., Gopalakrishnan, K., and Narayanan, P. (2015, January 6\u201311). Deep learning with limited numerical precision. Proceedings of the International Conference on Machine Learning, Lille, France."},{"key":"ref_61","unstructured":"Han, S., Mao, H., and Dally, W.J. (2016, January 2\u20134). Deep compression: Compressing deep neural networks with pruning, trained quantization and Huffman coding. Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico."},{"key":"ref_62","unstructured":"Ullrich, K., Meeds, E., and Welling, M. (2017, January 24\u201326). Soft weight-sharing for neural network compression. Proceedings of the International Conference on Learning Representations, Toulon, France."},{"key":"ref_63","unstructured":"Chen, W., Wilson, J., Tyree, S., Weinberger, K., and Chen, Y. (2015, January 6\u201311). Compressing neural networks with the hashing trick. Proceedings of the International Conference on Machine Learning, Lille, France."},{"key":"ref_64","unstructured":"Courbariaux, M., Hubara, I., Soudry, D., El-Yaniv, R., and Bengio, Y. (2016). Binarized neural networks: Training deep neural networks with weights and activations constrained to +1 or \u22121. arXiv."},{"key":"ref_65","unstructured":"Courbariaux, M., Bengio, Y., and David, J.P. (2015, January 7\u201312). Binaryconnect: Training deep neural networks with binary weights during propagations. Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada."},{"key":"ref_66","unstructured":"Hou, L., Yao, Q., and Kwok, J.T. (2017, January 24\u201326). Loss-aware binarization of deep networks. Proceedings of the International Conference on Learning Representations, Toulon, France."},{"key":"ref_67","unstructured":"Lin, Z., Courbariaux, M., Memisevic, R., and Bengio, Y. (2016, January 2\u20134). Neural networks with few multiplications. Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico."},{"key":"ref_68","unstructured":"Sironi, A., Tekin, B., Rigamonti, R., Lepetit, V., and Fua, P. (2013, January 23\u201328). Learning Separable Filters. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA."},{"key":"ref_69","doi-asserted-by":"crossref","unstructured":"Jaderberg, M., Vedaldi, A., and Zisserman, A. (2014, January 1\u20135). Speeding up Convolutional Neural Networks with Low Rank Expansions. Proceedings of the British Machine Vision Conference, Nottingham, UK.","DOI":"10.5244\/C.28.88"},{"key":"ref_70","doi-asserted-by":"crossref","unstructured":"Sainath, T., Kingsbury, B., Sindhwani, V., Arisoy, E., and Ramabhadran, B. (2013, January 22\u201327). Low-rank matrix factorization for Deep Neural Network training with high-dimensional output targets. Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Prague, Czech Republic.","DOI":"10.1109\/ICASSP.2013.6638949"},{"key":"ref_71","doi-asserted-by":"crossref","unstructured":"Lu, Y., Kumar, A., Zhai, S., Cheng, Y., Javidi, T., and Feris, R. (2017, January 21\u201326). Fully-Adaptive Feature Sharing in Multi-Task Networks with Applications in Person Attribute Classification. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.126"},{"key":"ref_72","unstructured":"Lebedev, V., Ganin, Y., Rakhuba, M., Oseledets, I., and Lempitsky, V. (2015, January 7\u20139). Speeding-up Convolutional Neural Networks Using Fine-tuned CP-Decomposition. Proceedings of the International Conference on Learning Representations, San Diego, CA, USA."},{"key":"ref_73","unstructured":"Tai, C., Xiao, T., Wang, X., and Weinan, E. (2016, January 2\u20134). Convolutional neural networks with low-rank regularization. Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico."},{"key":"ref_74","doi-asserted-by":"crossref","first-page":"126","DOI":"10.1109\/MSP.2017.2765695","article-title":"Model compression and acceleration for deep neural networks: The principles, progress, and challenges","volume":"35","author":"Cheng","year":"2018","journal-title":"IEEE Signal Process. Mag."},{"key":"ref_75","doi-asserted-by":"crossref","first-page":"359","DOI":"10.1007\/s12559-017-9506-0","article-title":"Compressing and accelerating neural network for facial point localization","volume":"10","author":"Zeng","year":"2018","journal-title":"Cogn. Comput."},{"key":"ref_76","doi-asserted-by":"crossref","unstructured":"Ge, S. (2018). Efficient deep learning in network compression and acceleration. Digital Systems, IntechOpen.","DOI":"10.5772\/intechopen.79562"}],"container-title":["Informatics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2227-9709\/8\/4\/77\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T07:31:25Z","timestamp":1760167885000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2227-9709\/8\/4\/77"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,11,17]]},"references-count":76,"journal-issue":{"issue":"4","published-online":{"date-parts":[[2021,12]]}},"alternative-id":["informatics8040077"],"URL":"https:\/\/doi.org\/10.3390\/informatics8040077","relation":{},"ISSN":["2227-9709"],"issn-type":[{"value":"2227-9709","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,11,17]]}}}