{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T22:31:43Z","timestamp":1770330703449,"version":"3.49.0"},"reference-count":63,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T00:00:00Z","timestamp":1732924800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T00:00:00Z","timestamp":1732924800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["No.U2133218"],"award-info":[{"award-number":["No.U2133218"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"the National Key Research and Development Program of China","award":["No.2018YFB0204304"],"award-info":[{"award-number":["No.2018YFB0204304"]}]},{"name":"the Fundamental Research Funds for the Central Universities of China","award":["No.FRF-MP-19-007"],"award-info":[{"award-number":["No.FRF-MP-19-007"]}]},{"name":"the Fundamental Research Funds for the Central Universities of China","award":["No.FRF-TP-20-065A1Z"],"award-info":[{"award-number":["No.FRF-TP-20-065A1Z"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1007\/s00530-024-01571-x","type":"journal-article","created":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T02:39:22Z","timestamp":1732934362000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["A systematic network pruning framework using ADMM-type algorithm"],"prefix":"10.1007","volume":"30","author":[{"given":"Fengrui","family":"Ji","sequence":"first","affiliation":[]},{"given":"Baolin","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,30]]},"reference":[{"issue":"6","key":"1571_CR1","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Commun. ACM 60(6), 84\u201390 (2017)","journal-title":"Commun. ACM"},{"issue":"ARTICLE","key":"1571_CR2","first-page":"2493","volume":"12","author":"R Collobert","year":"2011","unstructured":"Collobert, R., Weston, J., Bottou, L., Karlen, M., Kavukcuoglu, K., Kuksa, P.: Natural language processing (almost) from scratch. J. Mach. Learn. Res. 12(ARTICLE), 2493\u20132537 (2011)","journal-title":"J. Mach. Learn. Res."},{"issue":"5\u20136","key":"1571_CR3","doi-asserted-by":"publisher","first-page":"602","DOI":"10.1016\/j.neunet.2005.06.042","volume":"18","author":"A Graves","year":"2005","unstructured":"Graves, A., Schmidhuber, J.: Framewise phoneme classification with bidirectional lstm and other neural network architectures. Neural Netw. 18(5\u20136), 602\u2013610 (2005)","journal-title":"Neural Netw."},{"key":"1571_CR4","doi-asserted-by":"crossref","unstructured":"Lin, M., Ji, R., Wang, Y., Zhang, Y., Zhang, B., Tian, Y., Shao, L.: Hrank: Filter pruning using high-rank feature map. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1529\u20131538 (2020)","DOI":"10.1109\/CVPR42600.2020.00160"},{"key":"1571_CR5","unstructured":"Han, S., Pool, J., Tran, J., Dally, W.: Learning both weights and connections for efficient neural network. In: Advances in neural information processing systems 28 (2015)"},{"key":"1571_CR6","doi-asserted-by":"crossref","unstructured":"He, Y., Liu, P., Wang, Z., Hu, Z., Yang, Y.: Filter pruning via geometric median for deep convolutional neural networks acceleration. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4340\u20134349 (2019)","DOI":"10.1109\/CVPR.2019.00447"},{"issue":"5","key":"1571_CR7","doi-asserted-by":"publisher","first-page":"5611","DOI":"10.1007\/s40747-023-01022-6","volume":"9","author":"H Cheng","year":"2023","unstructured":"Cheng, H., Wang, Z., Ma, L., Wei, Z., Alsaadi, F.E., Liu, X.: Differentiable channel pruning guided via attention mechanism: a novel neural network pruning approach. Complex Intell. Syst. 9(5), 5611\u20135624 (2023)","journal-title":"Complex Intell. Syst."},{"issue":"5","key":"1571_CR8","doi-asserted-by":"publisher","first-page":"5779","DOI":"10.1007\/s40747-023-01036-0","volume":"9","author":"R Wang","year":"2023","unstructured":"Wang, R., Wan, S., Zhang, W., Zhang, C., Li, Y., Xu, S., Zhang, L., Jin, X., Jiang, Z., Rao, Y.: Progressive multi-level distillation learning for pruning network. Complex Intell. Syst. 9(5), 5779\u20135791 (2023)","journal-title":"Complex Intell. Syst."},{"issue":"6","key":"1571_CR9","doi-asserted-by":"publisher","first-page":"6999","DOI":"10.1007\/s40747-023-01120-5","volume":"9","author":"T Wu","year":"2023","unstructured":"Wu, T., Song, C., Zeng, P.: Efficient federated learning on resource-constrained edge devices based on model pruning. Complex Intell. Syst. 9(6), 6999\u20137013 (2023)","journal-title":"Complex Intell. Syst."},{"key":"1571_CR10","doi-asserted-by":"crossref","unstructured":"Li, X., Chen, L., Gao, Z., Zhang, X., Wang, C., Chen, H.: Lasso regression based channel pruning for efficient object detection model. In: 2019 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB), pp. 1\u20135 (2019). IEEE","DOI":"10.1109\/BMSB47279.2019.8971889"},{"issue":"3","key":"1571_CR11","doi-asserted-by":"publisher","first-page":"1463","DOI":"10.1007\/s00530-023-01062-5","volume":"29","author":"S Huang","year":"2023","unstructured":"Huang, S., Li, D., Zhang, Z., Wu, Y., Tang, Y., Chen, X., Wu, Y.: Cslsep: an ensemble pruning algorithm based on clustering soft label and sorting for facial expression recognition. Multimedia Syst. 29(3), 1463\u20131479 (2023)","journal-title":"Multimedia Syst."},{"issue":"3","key":"1571_CR12","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s00530-024-01315-x","volume":"30","author":"L Xiong","year":"2024","unstructured":"Xiong, L., Chen, Q., Huang, J., Huang, X., Huang, P., Wei, S.: Students and teachers learning together: a robust training strategy for neural network pruning. Multimedia Syst. 30(3), 1\u201314 (2024)","journal-title":"Multimedia Syst."},{"key":"1571_CR13","unstructured":"Hubara, I., Courbariaux, M., Soudry, D., El-Yaniv, R., Bengio, Y.: Binarized neural networks. In: Advances in neural information processing systems 29 (2016)"},{"key":"1571_CR14","unstructured":"Han, S., Mao, H., Dally, W.J.: Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 (2015)"},{"key":"1571_CR15","doi-asserted-by":"crossref","unstructured":"Peng, B., Tan, W., Li, Z., Zhang, S., Xie, D., Pu, S.: Extreme network compression via filter group approximation. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 300\u2013316 (2018)","DOI":"10.1007\/978-3-030-01237-3_19"},{"key":"1571_CR16","unstructured":"Hinton, G., Vinyals, O., Dean, J., et al.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.025312(7) (2015)"},{"key":"1571_CR17","doi-asserted-by":"crossref","unstructured":"Chen, L., Chen, Y., Xi, J., Le, X.: Knowledge from the original network: restore a better pruned network with knowledge distillation. Complex Intell. Syst., 1\u201310 (2021)","DOI":"10.1007\/s40747-020-00248-y"},{"key":"1571_CR18","unstructured":"Louizos, C., Welling, M., Kingma, D.P.: Learning sparse neural networks through $$l_0$$ regularization. arXiv preprint arXiv:1712.01312 (2017)"},{"key":"1571_CR19","doi-asserted-by":"crossref","unstructured":"Liu, Z., Li, J., Shen, Z., Huang, G., Yan, S., Zhang, C.: Learning efficient convolutional networks through network slimming. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2736\u20132744 (2017)","DOI":"10.1109\/ICCV.2017.298"},{"key":"1571_CR20","doi-asserted-by":"crossref","unstructured":"Zhang, T., Ye, S., Zhang, K., Tang, J., Wen, W., Fardad, M., Wang, Y.: A systematic dnn weight pruning framework using alternating direction method of multipliers. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 184\u2013199 (2018)","DOI":"10.1007\/978-3-030-01237-3_12"},{"key":"1571_CR21","doi-asserted-by":"crossref","unstructured":"Ma, X., Lin, S., Ye, S., He, Z., Zhang, L., Yuan, G., Tan, S.H., Li, Z., Fan, D., Qian, X., et al.: Non-structured dnn weight pruning\u2013is it beneficial in any platform? IEEE Trans. Neural Netw. Learn. Syst. (2021)","DOI":"10.1109\/TNNLS.2021.3063265"},{"issue":"1","key":"1571_CR22","doi-asserted-by":"publisher","first-page":"337","DOI":"10.1137\/140990309","volume":"26","author":"M Hong","year":"2016","unstructured":"Hong, M., Luo, Z.-Q., Razaviyayn, M.: Convergence analysis of alternating direction method of multipliers for a family of nonconvex problems. SIAM J. Optim. 26(1), 337\u2013364 (2016)","journal-title":"SIAM J. Optim."},{"key":"1571_CR23","unstructured":"Liu, S., Chen, J., Chen, P.-Y., Hero, A.: Zeroth-order online alternating direction method of multipliers: Convergence analysis and applications. In: International Conference on Artificial Intelligence and Statistics, pp. 288\u2013297 (2018). PMLR"},{"key":"1571_CR24","doi-asserted-by":"crossref","unstructured":"Ren, A., Zhang, T., Ye, S., Li, J., Xu, W., Qian, X., Lin, X., Wang, Y.: Admm-nn: An algorithm-hardware co-design framework of dnns using alternating direction methods of multipliers. In: Proceedings of the Twenty-Fourth International Conference on Architectural Support for Programming Languages and Operating Systems, pp. 925\u2013938 (2019)","DOI":"10.1145\/3297858.3304076"},{"key":"1571_CR25","unstructured":"Guo, Y., Yao, A., Chen, Y.: Dynamic network surgery for efficient dnns. In: Advances in neural information processing systems 29 (2016)"},{"key":"1571_CR26","unstructured":"Zhu, M., Gupta, S.: To prune, or not to prune: exploring the efficacy of pruning for model compression. arXiv preprint arXiv:1710.01878 (2017)"},{"key":"1571_CR27","unstructured":"Han, S., Pool, J., Narang, S., Mao, H., Gong, E., Tang, S., Elsen, E., Vajda, P., Paluri, M., Tran, J., et al.: Dsd: Dense-sparse-dense training for deep neural networks. arXiv preprint arXiv:1607.04381 (2016)"},{"key":"1571_CR28","first-page":"20852","volume":"33","author":"B Bartoldson","year":"2020","unstructured":"Bartoldson, B., Morcos, A., Barbu, A., Erlebacher, G.: The generalization-stability tradeoff in neural network pruning. Adv. Neural. Inf. Process. Syst. 33, 20852\u201320864 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1571_CR29","unstructured":"Hassibi, B., Stork, D.: Second order derivatives for network pruning: optimal brain surgeon. In: Advances in neural information processing systems 5 (1992)"},{"key":"1571_CR30","unstructured":"LeCun, Y., Denker, J., Solla, S.: Optimal brain damage. In: Advances in neural information processing systems 2 (1989)"},{"key":"1571_CR31","first-page":"18098","volume":"33","author":"SP Singh","year":"2020","unstructured":"Singh, S.P., Alistarh, D.: Woodfisher: efficient second-order approximation for neural network compression. Adv. Neural. Inf. Process. Syst. 33, 18098\u201318109 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1571_CR32","unstructured":"Theis, L., Korshunova, I., Tejani, A., Husz\u00e1r, F.: Faster gaze prediction with dense networks and fisher pruning. arXiv preprint arXiv:1801.05787 (2018)"},{"issue":"1","key":"1571_CR33","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1016\/S0925-2312(96)00054-9","volume":"16","author":"L Prechelt","year":"1997","unstructured":"Prechelt, L.: Connection pruning with static and adaptive pruning schedules. Neurocomputing 16(1), 49\u201361 (1997)","journal-title":"Neurocomputing"},{"key":"1571_CR34","unstructured":"Wortsman, M., Farhadi, A., Rastegari, M.: Discovering neural wirings. In: Advances in Neural Information Processing Systems 32 (2019)"},{"key":"1571_CR35","unstructured":"Liu, T., Zenke, F.: Finding trainable sparse networks through neural tangent transfer. In: International Conference on Machine Learning, pp. 6336\u20136347 (2020). PMLR"},{"key":"1571_CR36","unstructured":"Jin, X., Yuan, X., Feng, J., Yan, S.: Training skinny deep neural networks with iterative hard thresholding methods. arXiv preprint arXiv:1607.05423 (2016)"},{"key":"1571_CR37","unstructured":"Shwartz-Ziv, R., Tishby, N.: Opening the black box of deep neural networks via information. arXiv preprint arXiv:1703.00810 (2017)"},{"key":"1571_CR38","unstructured":"Ding, X., Zhou, X., Guo, Y., Han, J., Liu, J., et al.: Global sparse momentum sgd for pruning very deep neural networks. In: Advances in Neural Information Processing Systems 32 (2019)"},{"issue":"241","key":"1571_CR39","first-page":"1","volume":"22","author":"T Hoefler","year":"2021","unstructured":"Hoefler, T., Alistarh, D., Ben-Nun, T., Dryden, N., Peste, A.: Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks. J. Mach. Learn. Res. 22(241), 1\u2013124 (2021)","journal-title":"J. Mach. Learn. Res."},{"key":"1571_CR40","unstructured":"You, H., Li, C., Xu, P., Fu, Y., Wang, Y., Chen, X., Baraniuk, R.G., Wang, Z., Lin, Y.: Drawing early-bird tickets: Towards more efficient training of deep networks. arXiv preprint arXiv:1909.11957 (2019)"},{"key":"1571_CR41","first-page":"252","volume":"1","author":"M Lis","year":"2019","unstructured":"Lis, M., Golub, M., Lemieux, G.: Full deep neural network training on a pruned weight budget. Proc. Mach. Learn. Syst. 1, 252\u2013263 (2019)","journal-title":"Proc. Mach. Learn. Syst."},{"key":"1571_CR42","doi-asserted-by":"crossref","unstructured":"Zheng, X., Ji, R., Tang, L., Zhang, B., Liu, J., Tian, Q.: Multinomial distribution learning for effective neural architecture search. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1304\u20131313 (2019)","DOI":"10.1109\/ICCV.2019.00139"},{"key":"1571_CR43","unstructured":"Evci, U., Gale, T., Menick, J., Castro, P.S., Elsen, E.: Rigging the lottery: Making all tickets winners. In: International Conference on Machine Learning, pp. 2943\u20132952 (2020). PMLR"},{"key":"1571_CR44","unstructured":"Frankle, M. Jonathan\u00a0Carbin: The lottery ticket hypothesis: Finding sparse, trainable neural networks. In: International Conference on Learning Representations (2018)"},{"key":"1571_CR45","unstructured":"Frankle, J., Dziugaite, G.K., Roy, D.M., Carbin, M.: Stabilizing the lottery ticket hypothesis. arXiv preprint arXiv:1903.01611 (2019)"},{"issue":"1","key":"1571_CR46","first-page":"1","volume":"3","author":"S Boyd","year":"2010","unstructured":"Boyd, S., Parikh, N., Chu, E., Peleato, B., Eckstein, J.: Distributed optimization and statistical learning via the alternating direction method of multipliers. Mach. Learn. 3(1), 1\u2013122 (2010)","journal-title":"Mach. Learn."},{"key":"1571_CR47","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1017\/CBO9780511804441","volume-title":"Convex Optimization","author":"S Boyd","year":"2004","unstructured":"Boyd, S., Boyd, S.P., Vandenberghe, L.: Convex Optimization, pp. 1\u2013474. Cambridge University Press (2004)"},{"key":"1571_CR48","unstructured":"Zhang, T.: Analysis of multi-stage convex relaxation for sparse regularization. J. Mach. Learn. Res. 11(3) (2010)"},{"key":"1571_CR49","volume-title":"Deep Learning","author":"I Goodfellow","year":"2016","unstructured":"Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press (2016)"},{"key":"1571_CR50","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"1571_CR51","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"1571_CR52","unstructured":"Siegel, J.W., Chen, J., Zhang, P., Xu, J.: Training sparse neural networks using compressed sensing. arXiv preprint arXiv:2008.09661 (2020)"},{"key":"1571_CR53","doi-asserted-by":"crossref","unstructured":"Srinivas, S., Kuzmin, A., Nagel, M., Baalen, M., Skliar, A., Blankevoort, T.: Cyclical pruning for sparse neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2762\u20132771 (2022)","DOI":"10.1109\/CVPRW56347.2022.00312"},{"key":"1571_CR54","unstructured":"Lin, T., Stich, S.U., Barba, L., Dmitriev, D., Jaggi, M.: Dynamic model pruning with feedback. arXiv preprint arXiv:2006.07253 (2020)"},{"key":"1571_CR55","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1016\/j.neucom.2021.04.022","volume":"451","author":"K Xu","year":"2021","unstructured":"Xu, K., Zhang, D., An, J., Liu, L., Liu, L., Wang, D.: Genexp: multi-objective pruning for deep neural network based on genetic algorithm. Neurocomputing 451, 81\u201394 (2021)","journal-title":"Neurocomputing"},{"key":"1571_CR56","unstructured":"Bai, Y., Wang, H., TAO, Z., Li, K., Fu, Y.: Dual lottery ticket hypothesis. In: International Conference on Learning Representations (2021)"},{"key":"1571_CR57","unstructured":"Liu, J., Xu, Z., Shi, R., Cheung, R.C., So, H.K.: Dynamic sparse training: Find efficient sparse network from scratch with trainable masked layers. arXiv preprint arXiv:2005.06870 (2020)"},{"key":"1571_CR58","doi-asserted-by":"crossref","unstructured":"Shang, Y., Duan, B., Zong, Z., Nie, L., Yan, Y.: Win the lottery ticket via fourier analysis: Frequencies guided network pruning. In: ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4203\u20134207 (2022). IEEE","DOI":"10.1109\/ICASSP43922.2022.9746892"},{"key":"1571_CR59","unstructured":"Wang, H., Qin, C., Zhang, Y., Fu, Y.: Neural pruning via growing regularization. In: International Conference on Learning Representations (ICLR) (2021)"},{"key":"1571_CR60","unstructured":"Deng, W., Zhang, X., Liang, F., Lin, G.: An adaptive empirical bayesian method for sparse deep learning. In: Advances in neural information processing systems 32 (2019)"},{"key":"1571_CR61","doi-asserted-by":"crossref","unstructured":"Resende\u00a0Oliveira, F.D., Batista, E.L.O., Seara, R.: On the compression of neural networks using \u21130-norm regularization and weight pruning. Neural Netw. 171, 343\u2013352 (2024)","DOI":"10.1016\/j.neunet.2023.12.019"},{"key":"1571_CR62","unstructured":"Lubana, E.S., Dick, R.P.: A gradient flow framework for analyzing network pruning. arXiv preprint arXiv:2009.11839 (2020)"},{"key":"1571_CR63","unstructured":"Alizadeh, M., Tailor, S.A., Zintgraf, L.M., Amersfoort, J., Farquhar, S., Lane, N.D., Gal, Y.: Prospect pruning: Finding trainable weights at initialization using meta-gradients. arXiv preprint arXiv:2202.08132 (2022)"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01571-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01571-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01571-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,16]],"date-time":"2024-12-16T09:22:14Z","timestamp":1734340934000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01571-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,30]]},"references-count":63,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["1571"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01571-x","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,30]]},"assertion":[{"value":"26 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 November 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 November 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"373"}}