{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T15:34:36Z","timestamp":1772120076854,"version":"3.50.1"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2023,11,17]],"date-time":"2023-11-17T00:00:00Z","timestamp":1700179200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,11,17]],"date-time":"2023-11-17T00:00:00Z","timestamp":1700179200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["72074202"],"award-info":[{"award-number":["72074202"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Key Project of the Department of Science and Technology of Guangxi Zhuang Autonomous Region of China","award":["Gui Science AA22068069"],"award-info":[{"award-number":["Gui Science AA22068069"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Memetic Comp."],"published-print":{"date-parts":[[2024,3]]},"DOI":"10.1007\/s12293-023-00402-2","type":"journal-article","created":{"date-parts":[[2023,11,17]],"date-time":"2023-11-17T14:02:53Z","timestamp":1700229773000},"page":"45-54","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["An optimization method for pruning rates of each layer in CNN based on the GA-SMSM"],"prefix":"10.1007","volume":"16","author":[{"given":"Xiaoyu","family":"Dong","sequence":"first","affiliation":[]},{"given":"Pinshuai","family":"Yan","sequence":"additional","affiliation":[]},{"given":"Mengfei","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Binqi","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yuantao","family":"Song","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,11,17]]},"reference":[{"issue":"4","key":"402_CR1","doi-asserted-by":"publisher","first-page":"1249","DOI":"10.3390\/s21041249","volume":"21","author":"BJ Abbaschian","year":"2021","unstructured":"Abbaschian BJ, Sierra-Sosa D, Elmaghraby A (2021) Deep learning techniques for speech emotion recognition, from databases to models. Sensors 21(4):1249","journal-title":"Sensors"},{"key":"402_CR2","unstructured":"Denil M, Shakibi B, Dinh L, et\u00a0al (2013) Predicting parameters in deep learning. Adv Neural Inf Process Syst 26"},{"key":"402_CR3","doi-asserted-by":"crossref","unstructured":"Dong X, Song M, Li B, et\u00a0al (2022) A neural network pruning method by the sum of matrix similarity measures. In: 2022 4th international academic exchange conference on science and technology innovation (IAECST). IEEE, pp 679\u2013684","DOI":"10.1109\/IAECST57965.2022.10062205"},{"issue":"10","key":"402_CR4","first-page":"2911","volume":"25","author":"JK Ge","year":"2008","unstructured":"Ge JK, Qiu YH, Wu CM et al (2008) Summary of genetic algorithms research. Appl Res Comput 25(10):2911\u20132916","journal-title":"Appl Res Comput"},{"key":"402_CR5","unstructured":"Han S, Pool J, Tran J, et\u00a0al (2015) Learning both weights and connections for efficient neural network. Adv Neural Inf Process Syst 28"},{"key":"402_CR6","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, et\u00a0al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"402_CR7","doi-asserted-by":"crossref","unstructured":"He Y, Zhang X, Sun J (2017) Channel pruning for accelerating very deep neural networks. In: Proceedings of the IEEE international conference on computer vision, pp 1389\u20131397","DOI":"10.1109\/ICCV.2017.155"},{"key":"402_CR8","doi-asserted-by":"crossref","unstructured":"He Y, Kang G, Dong X, et\u00a0al (2018) Soft filter pruning for accelerating deep convolutional neural networks. arXiv preprint arXiv:1808.06866","DOI":"10.24963\/ijcai.2018\/309"},{"key":"402_CR9","doi-asserted-by":"crossref","unstructured":"He Y, Lin J, Liu Z, et\u00a0al (2018) Amc: automl for model compression and acceleration on mobile devices. In: Proceedings of the European conference on computer vision (ECCV), pp 784\u2013800","DOI":"10.1007\/978-3-030-01234-2_48"},{"key":"402_CR10","doi-asserted-by":"crossref","unstructured":"He Y, Liu P, Wang Z, et\u00a0al (2019) Filter pruning via geometric median for deep convolutional neural networks acceleration. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 4340\u20134349","DOI":"10.1109\/CVPR.2019.00447"},{"key":"402_CR11","unstructured":"Hinton G, Vinyals O, Dean J, et\u00a0al (2015) Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531"},{"key":"402_CR12","unstructured":"Jaderberg M, Dalibard V, Osindero S, et\u00a0al (2017) Population based training of neural networks. arXiv preprint arXiv:1711.09846"},{"key":"402_CR13","doi-asserted-by":"publisher","first-page":"62","DOI":"10.1016\/j.swevo.2019.05.010","volume":"49","author":"FEF Junior","year":"2019","unstructured":"Junior FEF, Yen GG (2019) Particle swarm optimization of deep neural networks architectures for image classification. Swarm Evol Comput 49:62\u201374","journal-title":"Swarm Evol Comput"},{"issue":"1","key":"402_CR14","doi-asserted-by":"publisher","first-page":"161","DOI":"10.1587\/transinf.2021EDP7096","volume":"105","author":"K Kamma","year":"2022","unstructured":"Kamma K, Inoue S, Wada T (2022) Pruning ratio optimization with layer-wise pruning method for accelerating convolutional neural networks. IEICE Trans Inf Syst 105(1):161\u2013169","journal-title":"IEICE Trans Inf Syst"},{"key":"402_CR15","doi-asserted-by":"publisher","first-page":"443","DOI":"10.1016\/j.neucom.2021.05.103","volume":"470","author":"I Lauriola","year":"2022","unstructured":"Lauriola I, Lavelli A, Aiolli F (2022) An introduction to deep learning in natural language processing: models, techniques, and tools. Neurocomputing 470:443\u2013456","journal-title":"Neurocomputing"},{"key":"402_CR16","unstructured":"Lee N, Ajanthan T, Torr PH (2018) Snip: Single-shot network pruning based on connection sensitivity. arXiv preprint arXiv:1810.02340"},{"key":"402_CR17","unstructured":"Li H, Kadav A, Durdanovic I, et\u00a0al (2016) Pruning filters for efficient convnets. arXiv preprint arXiv:1608.08710"},{"key":"402_CR18","doi-asserted-by":"crossref","unstructured":"Li T, Li J, Liu Z, et\u00a0al (2020) Few sample knowledge distillation for efficient network compression. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14639\u201314647","DOI":"10.1109\/CVPR42600.2020.01465"},{"key":"402_CR19","doi-asserted-by":"crossref","unstructured":"Li Y, Adamczewski K, Li W, et\u00a0al (2022) Revisiting random channel pruning for neural network compression. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 191\u2013201","DOI":"10.1109\/CVPR52688.2022.00029"},{"key":"402_CR20","doi-asserted-by":"crossref","unstructured":"Lin S, Ji R, Chen C, et\u00a0al (2017) Espace: accelerating convolutional neural networks via eliminating spatial and channel redundancy. In: Thirty-first AAAI conference on artificial intelligence","DOI":"10.1609\/aaai.v31i1.10756"},{"issue":"7","key":"402_CR21","first-page":"1573","volume":"37","author":"H Liu","year":"2016","unstructured":"Liu H, Zhao C, Li X et al (2016) Study on a neural network optimization algorithm based on improved genetic algorithm. Chin J Sci Instrum 37(7):1573\u20131681","journal-title":"Chin J Sci Instrum"},{"key":"402_CR22","doi-asserted-by":"crossref","unstructured":"Liu Z, Li J, Shen Z, et\u00a0al (2017) Learning efficient convolutional networks through network slimming. In: Proceedings of the IEEE international conference on computer vision, pp 2736\u20132744","DOI":"10.1109\/ICCV.2017.298"},{"key":"402_CR23","doi-asserted-by":"crossref","unstructured":"Luo JH, Wu J, Lin W (2017) Thinet: a filter level pruning method for deep neural network compression. In: Proceedings of the IEEE international conference on computer vision, pp 5058\u20135066","DOI":"10.1109\/ICCV.2017.541"},{"key":"402_CR24","unstructured":"Molchanov P, Tyree S, Karras T, et\u00a0al (2016) Pruning convolutional neural networks for resource efficient inference. arXiv preprint arXiv:1611.06440"},{"key":"402_CR25","doi-asserted-by":"crossref","unstructured":"Molchanov P, Mallya A, Tyree S, et\u00a0al (2019) Importance estimation for neural network pruning. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 11264\u201311272","DOI":"10.1109\/CVPR.2019.01152"},{"key":"402_CR26","unstructured":"Neill JO (2020) An overview of neural network compression. arXiv preprint arXiv:2006.03669"},{"issue":"11","key":"402_CR27","first-page":"75","volume":"4","author":"KB Obaid","year":"2020","unstructured":"Obaid KB, Zeebaree S, Ahmed OM et al (2020) Deep learning models based on image classification: a review. Int J Sci Bus 4(11):75\u201381","journal-title":"Int J Sci Bus"},{"issue":"2","key":"402_CR28","doi-asserted-by":"publisher","first-page":"819","DOI":"10.1007\/s10044-020-00940-2","volume":"24","author":"SW Prakosa","year":"2021","unstructured":"Prakosa SW, Leu JS, Chen ZH (2021) Improving the accuracy of pruned network using knowledge distillation. Pattern Anal Appl 24(2):819\u2013830","journal-title":"Pattern Anal Appl"},{"issue":"9","key":"402_CR29","first-page":"1871","volume":"55","author":"J Rongrong","year":"2018","unstructured":"Rongrong J, Shaohui L, Fei C et al (2018) Deep neural network compression and acceleration: A review. J Comput Res Dev 55(9):1871","journal-title":"J Comput Res Dev"},{"key":"402_CR30","doi-asserted-by":"crossref","unstructured":"Sakai Y, Iwakawa A, Tabaru T, et\u00a0al (2022) Automatic pruning rate derivation for structured pruning of deep neural networks. In: 2022 26th international conference on pattern recognition (ICPR). IEEE, pp 2561\u20132567","DOI":"10.1109\/ICPR56361.2022.9956644"},{"key":"402_CR31","doi-asserted-by":"crossref","unstructured":"Sampson JR (1976) Adaptation in natural and artificial systems (John H. Holland)","DOI":"10.1137\/1018105"},{"issue":"2","key":"402_CR32","doi-asserted-by":"publisher","first-page":"381","DOI":"10.1007\/s11760-020-01760-x","volume":"15","author":"M Shao","year":"2021","unstructured":"Shao M, Dai J, Kuang J et al (2021) A dynamic cnn pruning method based on matrix similarity. SIViP 15(2):381\u2013389","journal-title":"SIViP"},{"key":"402_CR33","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556"},{"key":"402_CR34","doi-asserted-by":"publisher","DOI":"10.1016\/j.swevo.2021.100863","volume":"63","author":"P Singh","year":"2021","unstructured":"Singh P, Chaudhury S, Panigrahi BK (2021) Hybrid mpso-cnn: multi-level particle swarm optimized hyperparameters of convolutional neural network. Swarm Evol Comput 63:100863","journal-title":"Swarm Evol Comput"},{"key":"402_CR35","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1016\/j.neucom.2020.02.035","volume":"398","author":"S Swaminathan","year":"2020","unstructured":"Swaminathan S, Garg D, Kannan R et al (2020) Sparse low rank factorization for deep neural network compression. Neurocomputing 398:185\u2013196","journal-title":"Neurocomputing"},{"key":"402_CR36","unstructured":"Tan M, Le Q (2021) Efficientnetv2: smaller models and faster training. In: International conference on machine learning. PMLR, pp 10096\u201310106"},{"key":"402_CR37","doi-asserted-by":"publisher","first-page":"247","DOI":"10.1016\/j.neucom.2020.03.082","volume":"404","author":"Z Wang","year":"2020","unstructured":"Wang Z, Li F, Shi G et al (2020) Network pruning using sparse learning and genetic algorithm. Neurocomputing 404:247\u2013256","journal-title":"Neurocomputing"},{"key":"402_CR38","unstructured":"Wu J, Wang Y, Wu Z, et\u00a0al (2018) Deep k-means: re-training and parameter sharing with harder cluster assignments for compressing deep convolutions. In: International conference on machine learning. PMLR, pp 5363\u20135372"},{"key":"402_CR39","doi-asserted-by":"crossref","unstructured":"Yang C, An Z, Li C, et\u00a0al (2019) Multi-objective pruning for cnns using genetic algorithm. In: International conference on artificial neural networks. Springer, pp 299\u2013305","DOI":"10.1007\/978-3-030-30484-3_25"},{"key":"402_CR40","doi-asserted-by":"crossref","unstructured":"Yang W, Xiao Y (2022) Structured pruning via feature channels similarity and mutual learning for convolutional neural network compression. Appl Intell 1\u201311","DOI":"10.1007\/s10489-022-03403-9"},{"key":"402_CR41","unstructured":"You Z, Yan K, Ye J, et\u00a0al (2019) Gate decorator: global filter pruning method for accelerating deep convolutional neural networks. Adv Neural Inf Process Syst 32"},{"issue":"06","key":"402_CR42","first-page":"83","volume":"44","author":"YL Zhu","year":"2021","unstructured":"Zhu YL, Liu YM (2021) Hyper parameter optimization of convolutional neural network based on multi-strategy mrfo algorithm. J Beijing Univ Posts Telecommun 44(06):83","journal-title":"J Beijing Univ Posts Telecommun"}],"container-title":["Memetic Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12293-023-00402-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s12293-023-00402-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12293-023-00402-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T04:48:40Z","timestamp":1710305320000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s12293-023-00402-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,17]]},"references-count":42,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2024,3]]}},"alternative-id":["402"],"URL":"https:\/\/doi.org\/10.1007\/s12293-023-00402-2","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-2738666\/v1","asserted-by":"object"}]},"ISSN":["1865-9284","1865-9292"],"issn-type":[{"value":"1865-9284","type":"print"},{"value":"1865-9292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,11,17]]},"assertion":[{"value":"26 March 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 November 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 November 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This work was supported by the National Natural Science Foundation of China under Grant 72074202 and the Key Project of the Science and Technology Department of Guangxi Zhuang Autonomous of China under Grant Gui Science AA22068069. The authors have no relevant financial or non-financial interests to disclose.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"Not applicable.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}]}}