{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T15:21:52Z","timestamp":1772119312545,"version":"3.50.1"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2025,3,13]],"date-time":"2025-03-13T00:00:00Z","timestamp":1741824000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,3,13]],"date-time":"2025-03-13T00:00:00Z","timestamp":1741824000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No.U2133218"],"award-info":[{"award-number":["No.U2133218"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["No.FRF-MP-19-007 and No.FRF-TP-20-065A1Z"],"award-info":[{"award-number":["No.FRF-MP-19-007 and No.FRF-TP-20-065A1Z"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]},{"name":"National Key Research and Development Program of China","award":["No.2024YFC3308300"],"award-info":[{"award-number":["No.2024YFC3308300"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2025,6]]},"DOI":"10.1007\/s00530-025-01743-3","type":"journal-article","created":{"date-parts":[[2025,3,13]],"date-time":"2025-03-13T07:53:38Z","timestamp":1741852418000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Data-free pruning of CNN using kernel similarity"],"prefix":"10.1007","volume":"31","author":[{"given":"Xinwang","family":"Chen","sequence":"first","affiliation":[]},{"given":"Fengrui","family":"Ji","sequence":"additional","affiliation":[]},{"given":"Renxin","family":"Chu","sequence":"additional","affiliation":[]},{"given":"Baolin","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,3,13]]},"reference":[{"issue":"6","key":"1743_CR1","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Commun ACM 60(6), 84\u201390 (2017)","journal-title":"Commun ACM"},{"key":"1743_CR2","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3431\u20133440 (2015)","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"1743_CR3","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"1743_CR4","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Adv. Neural Inform. Process. Syst. 33, 6840\u20136851 (2020)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"1743_CR5","unstructured":"LeCun, Y., Denker, J., Solla, S.: Optimal brain damage. Adv. Neural Inform. Process. Syst. 2 (1989)"},{"key":"1743_CR6","doi-asserted-by":"crossref","unstructured":"Liu, N., Ma, X., Xu, Z., Wang, Y., Tang, J., Ye, J.: Autocompress: An automatic dnn structured pruning framework for ultra-high compression rates. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 4876\u20134883 (2020)","DOI":"10.1609\/aaai.v34i04.5924"},{"key":"1743_CR7","first-page":"129","volume":"2","author":"D Blalock","year":"2020","unstructured":"Blalock, D., Gonzalez Ortiz, J.J., Frankle, J., Guttag, J.: What is the state of neural network pruning? Proc. Mach. Learn. Syst. 2, 129\u2013146 (2020)","journal-title":"Proc. Mach. Learn. Syst."},{"issue":"6","key":"1743_CR8","doi-asserted-by":"publisher","first-page":"373","DOI":"10.1007\/s00530-024-01571-x","volume":"30","author":"F Ji","year":"2024","unstructured":"Ji, F., Liu, B.: A systematic network pruning framework using admm-type algorithm. Multimed. Syst. 30(6), 373 (2024)","journal-title":"Multimed. Syst."},{"issue":"3","key":"1743_CR9","doi-asserted-by":"publisher","first-page":"1463","DOI":"10.1007\/s00530-023-01062-5","volume":"29","author":"S Huang","year":"2023","unstructured":"Huang, S., Li, D., Zhang, Z., Wu, Y., Tang, Y., Chen, X., Wu, Y.: Cslsep: an ensemble pruning algorithm based on clustering soft label and sorting for facial expression recognition. Multimed. Syst. 29(3), 1463\u20131479 (2023)","journal-title":"Multimed. Syst."},{"issue":"1","key":"1743_CR10","doi-asserted-by":"publisher","first-page":"46","DOI":"10.1007\/s00530-023-01227-2","volume":"30","author":"Y Wu","year":"2024","unstructured":"Wu, Y., Li, D., Chen, X., Tang, Y., Huang, S.: An ensemble pruning method considering classifiers\u2019 interaction based on information theory for facial expression recognition. Multimed. Syst. 30(1), 46 (2024)","journal-title":"Multimed. Syst."},{"key":"1743_CR11","unstructured":"Li, Y., Yu, Y., Zhang, Q., Liang, C., He, P., Chen, W., Zhao, T.: Losparse: Structured compression of large language models based on low-rank and sparse approximation. In: International Conference on Machine Learning, pp. 20336\u201320350. PMLR (2023)"},{"key":"1743_CR12","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1016\/j.neucom.2020.02.035","volume":"398","author":"S Swaminathan","year":"2020","unstructured":"Swaminathan, S., Garg, D., Kannan, R., Andres, F.: Sparse low rank factorization for deep neural network compression. Neurocomputing 398, 185\u2013196 (2020)","journal-title":"Neurocomputing"},{"key":"1743_CR13","unstructured":"Denton, E.L., Zaremba, W., Bruna, J., LeCun, Y., Fergus, R.: Exploiting linear structure within convolutional networks for efficient evaluation. Adv. Neural Inform. Process. Syst. 27 (2014)"},{"key":"1743_CR14","doi-asserted-by":"crossref","unstructured":"Jacob, B., Kligys, S., Chen, B., Zhu, M., Tang, M., Howard, A., Adam, H., Kalenichenko, D.: Quantization and training of neural networks for efficient integer-arithmetic-only inference. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2704\u20132713 (2018)","DOI":"10.1109\/CVPR.2018.00286"},{"key":"1743_CR15","unstructured":"Courbariaux, M., Hubara, I., Soudry, D., El-Yaniv, R., Bengio, Y.: Binarized neural networks: training deep neural networks with weights and activations constrained to+ 1 or-1 (2016). arXiv preprint arXiv:1602.02830"},{"issue":"5","key":"1743_CR16","doi-asserted-by":"publisher","first-page":"696","DOI":"10.1109\/TC.2020.2995593","volume":"70","author":"C Gong","year":"2020","unstructured":"Gong, C., Chen, Y., Lu, Y., Li, T., Hao, C., Chen, D.: Vecq: minimal loss dnn model compression with vectorized weight quantization. IEEE Trans. Comput. 70(5), 696\u2013710 (2020)","journal-title":"IEEE Trans. Comput."},{"key":"1743_CR17","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network (2015). arXiv preprint arXiv:1503.02531"},{"key":"1743_CR18","unstructured":"Liang, C., Zuo, S., Zhang, Q., He, P., Chen, W., Zhao, T.: Less is more: task-aware layer-wise distillation for language model compression. In: International Conference on Machine Learning, pp. 20852\u201320867. PMLR (2023)"},{"key":"1743_CR19","unstructured":"Pan, H., Wang, C., Qiu, M., Zhang, Y., Li, Y., Huang, J.: Meta-kd: A meta knowledge distillation framework for language model compression across domains (2020). arXiv preprint arXiv:2012.01266"},{"key":"1743_CR20","doi-asserted-by":"crossref","unstructured":"Liu, Z., Li, J., Shen, Z., Huang, G., Yan, S., Zhang, C.: Learning efficient convolutional networks through network slimming. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2736\u20132744 (2017)","DOI":"10.1109\/ICCV.2017.298"},{"key":"1743_CR21","doi-asserted-by":"crossref","unstructured":"Lin, M., Ji, R., Wang, Y., Zhang, Y., Zhang, B., Tian, Y., Shao, L.: Hrank: filter pruning using high-rank feature map. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1529\u20131538 (2020)","DOI":"10.1109\/CVPR42600.2020.00160"},{"key":"1743_CR22","doi-asserted-by":"crossref","unstructured":"Guo, S., Zhang, L., Zheng, X., Wang, Y., Li, Y., Chao, F., Wu, C., Zhang, S., Ji, R.: Automatic network pruning via hilbert-schmidt independence criterion lasso under information bottleneck principle. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 17458\u201317469 (2023)","DOI":"10.1109\/ICCV51070.2023.01601"},{"key":"1743_CR23","first-page":"20863","volume":"34","author":"E Yvinec","year":"2021","unstructured":"Yvinec, E., Dapogny, A., Cord, M., Bailly, K.: Red: looking for redundancies for data-freestructured compression of deep neural networks. Adv. Neural Inform. Process. Syst. 34, 20863\u201320873 (2021)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"1743_CR24","doi-asserted-by":"crossref","unstructured":"Bai, S., Chen, J., Shen, X., Qian, Y., Liu, Y.: Unified data-free compression: Pruning and quantization without fine-tuning. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5876\u20135885 (2023)","DOI":"10.1109\/ICCV51070.2023.00540"},{"key":"1743_CR25","unstructured":"Li, S., Chen, J., Xiang, J., Zhu, C., Liu, Y.: Autodfp: Automatic data-free pruning via channel similarity reconstruction (2024). arXiv preprint arXiv:2403.08204"},{"key":"1743_CR26","unstructured":"Wen, W., Wu, C., Wang, Y., Chen, Y., Li, H.: Learning structured sparsity in deep neural networks. Adv. Neural Inform. Process. Syst. 29 (2016)"},{"issue":"14","key":"1743_CR27","doi-asserted-by":"publisher","first-page":"17429","DOI":"10.1007\/s10489-022-04353-y","volume":"53","author":"G Bonetta","year":"2023","unstructured":"Bonetta, G., Ribero, M., Cancelliere, R.: Regularization-based pruning of irrelevant weights in deep neural architectures. Appl. Intell. 53(14), 17429\u201317443 (2023)","journal-title":"Appl. Intell."},{"issue":"3","key":"1743_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s00530-024-01315-x","volume":"30","author":"L Xiong","year":"2024","unstructured":"Xiong, L., Chen, Q., Huang, J., Huang, X., Huang, P., Wei, S.: Students and teachers learning together: a robust training strategy for neural network pruning. Multimed. Syst. 30(3), 1\u201314 (2024)","journal-title":"Multimed. Syst."},{"key":"1743_CR29","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2023.103682","volume":"231","author":"Q Tian","year":"2023","unstructured":"Tian, Q., Arbel, T., Clark, J.J.: Grow-push-prune: aligning deep discriminants for effective structural network compression. Comput. Vis. Image Understand. 231, 103682 (2023)","journal-title":"Comput. Vis. Image Understand."},{"key":"1743_CR30","unstructured":"Frankle, J., Carbin, M.: The lottery ticket hypothesis: Finding sparse, trainable neural networks (2018). arXiv preprint arXiv:1803.03635"},{"key":"1743_CR31","unstructured":"Chen, T., Chen, X., Ma, X., Wang, Y., Wang, Z.: Coarsening the granularity: towards structurally sparse lottery tickets. In: International Conference on Machine Learning, pp. 3025\u20133039. PMLR (2022)"},{"key":"1743_CR32","doi-asserted-by":"crossref","unstructured":"Li, F., Li, G., He, X., Cheng, J.: Dynamic dual gating neural networks. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5330\u20135339 (2021)","DOI":"10.1109\/ICCV48922.2021.00528"},{"key":"1743_CR33","doi-asserted-by":"crossref","unstructured":"Tang, Y., You, S., Xu, C., Han, J., Qian, C., Shi, B., Xu, C., Zhang, C.: Reborn filters: pruning convolutional neural networks with limited data. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 5972\u20135980 (2020)","DOI":"10.1609\/aaai.v34i04.6058"},{"key":"1743_CR34","doi-asserted-by":"crossref","unstructured":"Tang, J., Liu, M., Jiang, N., Cai, H., Yu, W., Zhou, J.: Data-free network pruning for model compression. In: 2021 IEEE International Symposium on Circuits and Systems (ISCAS), pp. 1\u20135. IEEE (2021)","DOI":"10.1109\/ISCAS51556.2021.9401109"},{"key":"1743_CR35","first-page":"6377","volume":"33","author":"H Tanaka","year":"2020","unstructured":"Tanaka, H., Kunin, D., Yamins, D.L., Ganguli, S.: Pruning neural networks without any data by iteratively conserving synaptic flow. Adv. Neural Inform. Process. Syst. 33, 6377\u20136389 (2020)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"1743_CR36","unstructured":"Pei, Z., Zhang, A., Wang, S., Ji, X., Huang, Q.: Data-free neural representation compression with riemannian neural dynamics. In: Forty-first International Conference on Machine Learning"},{"key":"1743_CR37","first-page":"9865","volume":"33","author":"T Zhuang","year":"2020","unstructured":"Zhuang, T., Zhang, Z., Huang, Y., Zeng, X., Shuang, K., Li, X.: Neuron-level structured pruning using polarization regularizer. Adv. Neural Inform. Process. Syst. 33, 9865\u20139877 (2020)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"1743_CR38","first-page":"10936","volume":"33","author":"Y Tang","year":"2020","unstructured":"Tang, Y., Wang, Y., Xu, Y., Tao, D., Xu, C., Xu, C., Xu, C.: Scop: Scientific control for reliable neural network pruning. Adv. Neural Inform. Process. Syst. 33, 10936\u201310947 (2020)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"1743_CR39","first-page":"585","volume":"33","author":"W Kim","year":"2020","unstructured":"Kim, W., Kim, S., Park, M., Jeon, G.: Neuron merging: compensating for pruned neurons. Adv. Neural Inform. Process. Syst. 33, 585\u2013595 (2020)","journal-title":"Adv. Neural Inform. Process. Syst."},{"key":"1743_CR40","doi-asserted-by":"crossref","unstructured":"Wu, J., Zhu, D., Fang, L., Deng, Y., Zhong, Z.: Efficient layer compression without pruning. IEEE Trans. Image Process. (2023)","DOI":"10.1109\/TIP.2023.3302519"},{"key":"1743_CR41","doi-asserted-by":"crossref","unstructured":"Xiao, J., Li, P., Tang, Z., Nie, J.: Lnpt: label-free network pruning and training. In: International Joint Conference on Neural Networks, IJCNN 2024, Yokohama, Japan, June 30 - July 5, 2024, pp. 1\u20138. IEEE (2024)","DOI":"10.1109\/IJCNN60899.2024.10651544"},{"issue":"8","key":"1743_CR42","doi-asserted-by":"publisher","first-page":"3435","DOI":"10.1007\/s10115-023-01846-1","volume":"65","author":"Y Gao","year":"2023","unstructured":"Gao, Y., Zhang, Z., Zhang, H., Zhao, M., Yang, Y., Wang, M.: Fast data-free model compression via dictionary-pair reconstruction. Knowl. Inform. Syst. 65(8), 3435\u20133461 (2023)","journal-title":"Knowl. Inform. Syst."},{"key":"1743_CR43","doi-asserted-by":"crossref","unstructured":"Li, Y., Lin, S., Liu, J., Ye, Q., Wang, M., Chao, F., Yang, F., Ma, J., Tian, Q., Ji, R.: Towards compact cnns via collaborative compression. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6438\u20136447 (2021)","DOI":"10.1109\/CVPR46437.2021.00637"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01743-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-01743-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01743-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,4]],"date-time":"2025-09-04T15:00:44Z","timestamp":1756998044000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-01743-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,3,13]]},"references-count":43,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,6]]}},"alternative-id":["1743"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-01743-3","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-4919297\/v1","asserted-by":"object"}]},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,3,13]]},"assertion":[{"value":"15 August 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 February 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 March 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This research has no potential Conflict of interest, which encompass financial, non-financial, or other associations with individuals or organizations that could improperly impact our work.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The data used in this study is legally obtained from public datasets. It has been acquired with proper permissions and authorizations, ensuring compliance with ethical standards.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical and informed consent for data used"}}],"article-number":"154"}}