{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,16]],"date-time":"2025-10-16T07:02:37Z","timestamp":1760598157456,"version":"3.37.3"},"reference-count":38,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2022,7,14]],"date-time":"2022-07-14T00:00:00Z","timestamp":1657756800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,7,14]],"date-time":"2022-07-14T00:00:00Z","timestamp":1657756800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61671099,62176051"],"award-info":[{"award-number":["61671099,62176051"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Process Lett"],"published-print":{"date-parts":[[2023,4]]},"DOI":"10.1007\/s11063-022-10956-w","type":"journal-article","created":{"date-parts":[[2022,7,14]],"date-time":"2022-07-14T19:31:16Z","timestamp":1657827076000},"page":"1663-1679","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Batch Gradient Training Method with Smoothing Group $$L_0$$ Regularization for Feedfoward Neural Networks"],"prefix":"10.1007","volume":"55","author":[{"given":"Ying","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Jianing","family":"Wei","sequence":"additional","affiliation":[]},{"given":"Dongpo","family":"Xu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7394-7295","authenticated-orcid":false,"given":"Huisheng","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,7,14]]},"reference":[{"issue":"4","key":"10956_CR1","doi-asserted-by":"publisher","first-page":"659","DOI":"10.1109\/TKDE.2019.2893266","volume":"32","author":"H Zhang","year":"2020","unstructured":"Zhang H, Wang J, Sun Z, Zurada JM, Pal NR (2020) Feature selection for neural networks using Group Lasso regularization. IEEE Trans Knowl Data Eng 32(4):659\u2013673","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"10956_CR2","doi-asserted-by":"crossref","unstructured":"Liu T, Xiao J, Huang Z, Kong E, Liang Y (2019) BP neural network feature selection based on Group Lasso regularization. Proc. Chin. Autom. Congr. 2786-2790","DOI":"10.1109\/CAC48633.2019.8996679"},{"key":"10956_CR3","doi-asserted-by":"publisher","first-page":"9540","DOI":"10.1109\/ACCESS.2018.2890740","volume":"7","author":"HZ Alemu","year":"2019","unstructured":"Alemu HZ, Zhao J, Li F, Wu W (2019) Group $$L_{1\/2}$$ regularization for pruning hidden layer nodes of feedforward neural networks. IEEE Access 7:9540\u20139557","journal-title":"IEEE Access"},{"key":"10956_CR4","doi-asserted-by":"publisher","first-page":"241","DOI":"10.1007\/s11063-011-9196-7","volume":"34","author":"MG Augasta","year":"2011","unstructured":"Augasta MG, Kathirvalavakumar T (2011) A novel pruning algorithm for optimizing feedforward neural network of classifification problems. Neural Process Lett 34:241\u2013258","journal-title":"Neural Process Lett"},{"key":"10956_CR5","doi-asserted-by":"publisher","first-page":"825","DOI":"10.1016\/j.neucom.2005.04.010","volume":"69","author":"XQ Zeng","year":"2006","unstructured":"Zeng XQ, Yeung DS (2006) Hidden neuron pruning of multilayer perceptrons using a quantified sensitivity measure. Neurocomputing 69:825\u2013837","journal-title":"Neurocomputing"},{"key":"10956_CR6","doi-asserted-by":"publisher","first-page":"4346","DOI":"10.1109\/TCYB.2018.2864142","volume":"49","author":"J Wang","year":"2019","unstructured":"Wang J, Chang Q, Chang Q, Liu Y, Pal NR (2019) Weight noise injection-based MLPs with Group Lasso penalty: asymptotic convergence and application to node pruning. IEEE T Cybern 49:4346\u20134364","journal-title":"IEEE T Cybern"},{"issue":"5","key":"10956_CR7","doi-asserted-by":"publisher","first-page":"2012","DOI":"10.1109\/TNNLS.2017.2748585","volume":"29","author":"J Wang","year":"2018","unstructured":"Wang J, Xu C, Yang X, Zurada JM (2018) A novel pruning algorithm for smoothing feedforward neural networks based on Group Lasso method. IEEE Trans Neural Netw Learn Syst 29(5):2012\u20132024","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"10956_CR8","unstructured":"Dheeru D, Taniskidou EK (2017) UCI machine learning repository. Comput. Sci. Univ. California, Irvine, CA, USA, Tech. Rep, School Inf"},{"key":"10956_CR9","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1109\/72.478388","volume":"7","author":"JO Moody","year":"1996","unstructured":"Moody JO, Antsaklis PJ (1996) The dependence identification neural network construction algorithm. IEEE Trans Neural Netw 7:3\u201315","journal-title":"IEEE Trans Neural Netw"},{"key":"10956_CR10","first-page":"105","volume":"3","author":"MG Augasta","year":"2013","unstructured":"Augasta MG, Kathirvalavakumar T (2013) Pruning algorithms of neural networks a comparative study. Central Eur J Comput Sci 3:105\u2013115","journal-title":"Central Eur J Comput Sci"},{"key":"10956_CR11","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1109\/72.248452","volume":"4","author":"R Reed","year":"1993","unstructured":"Reed R (1993) Pruning algorithms: a survey. IEEE Trans Neural Netw 4:740\u2013747","journal-title":"IEEE Trans Neural Netw"},{"key":"10956_CR12","doi-asserted-by":"publisher","first-page":"796","DOI":"10.1016\/j.neucom.2020.03.119","volume":"452","author":"XY Wang","year":"2021","unstructured":"Wang XY, Wang J, Zhang K, Lin F, Chang Q (2021) Convergence and objective functions of noise-injected multilayer perceptrons with hidden multipliers. Neurocomputing 452:796\u2013812","journal-title":"Neurocomputing"},{"issue":"7","key":"10956_CR13","doi-asserted-by":"publisher","first-page":"1013","DOI":"10.1109\/TNNLS.2012.2197412","volume":"23","author":"ZB Xu","year":"2012","unstructured":"Xu ZB, Chang XY, Xu FM, Zhang H (2012) $$L_{1\/2}$$ regularization: a thresholding representation theory and a fast solver. IEEE Trans Neural Netw Learn Syst 23(7):1013\u201327","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"10956_CR14","doi-asserted-by":"publisher","first-page":"4355","DOI":"10.1109\/ACCESS.2016.2596704","volume":"4","author":"C Miao","year":"2016","unstructured":"Miao C, Yu H (2016) Alternating iteration for $$L_p(0 < p)$$ regularized CT reconstruction. IEEE Access 4:4355\u20134363","journal-title":"IEEE Access"},{"issue":"4","key":"10956_CR15","doi-asserted-by":"publisher","first-page":"662","DOI":"10.1109\/72.701179","volume":"9","author":"NK Treadgold","year":"1998","unstructured":"Treadgold NK, Gedeon TD (1998) Simulated annealing and weight decay in adaptive learning: the SARPROP algorithm. IEEE Trans Neural Netw 9(4):662\u20138","journal-title":"IEEE Trans Neural Netw"},{"key":"10956_CR16","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.neunet.2013.11.006","volume":"50","author":"W Wu","year":"2014","unstructured":"Wu W, Fan Q, Zurada JM, Wang J, Yang D, Liu Y (2014) Batch gradient method with smoothing $$L_{1\/2}$$ regularization for training of feedforward neural networks. Neural Netw 50:72\u201378","journal-title":"Neural Netw"},{"key":"10956_CR17","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1016\/j.neucom.2020.04.114","volume":"407","author":"HS Zhang","year":"2020","unstructured":"Zhang HS, Zhang Y, Zhu S, Xu DP (2020) Deterministic convergence of complex mini-batch gradient learning algorithm for fully complex-valued neural networks. Neurocomputing 407:185\u2013193","journal-title":"Neurocomputing"},{"issue":"1","key":"10956_CR18","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1016\/j.neunet.2010.09.007","volume":"24","author":"W Wu","year":"2011","unstructured":"Wu W, Wang J, Cheng MS, Li ZX (2011) Convergence analysis of online gradient method for BP neural networks. Neural Netw 24(1):91\u20138","journal-title":"Neural Netw"},{"key":"10956_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.neucom.2016.10.057","volume":"224","author":"HS Zhang","year":"2017","unstructured":"Zhang HS, Tang YL (2017) Online gradient method with smoothing $$L_0$$ regularization for feedforward neural networks. Neurocomputing 224:1\u20138","journal-title":"Neurocomputing"},{"key":"10956_CR20","doi-asserted-by":"publisher","first-page":"2368","DOI":"10.1016\/j.neucom.2011.03.016","volume":"74","author":"J Wang","year":"2011","unstructured":"Wang J, Wu W, Zurada JM (2011) Deterministic convergence of conjugate gradient mehtod for feedforward neural networks. Neurocomputing 74:2368\u20132376","journal-title":"Neurocomputing"},{"key":"10956_CR21","doi-asserted-by":"publisher","first-page":"141","DOI":"10.1016\/j.neucom.2012.02.029","volume":"89","author":"HS Zhang","year":"2012","unstructured":"Zhang HS, Wu W, Yao MC (2012) Boundedness and convergence of batch back-propagation algorithm with penalty for feedforward neural networks. Neurocomputing 89:141\u2013146","journal-title":"Neurocomputing"},{"key":"10956_CR22","doi-asserted-by":"publisher","first-page":"122","DOI":"10.1016\/j.neucom.2017.06.061","volume":"272","author":"DK Yang","year":"2018","unstructured":"Yang DK, Liu Y (2018) $$L_{1\/2}$$ regularization learning for smoothing interval neural networks: Algorithms and convergence analysis. Neurocomputing 272:122\u2013129","journal-title":"Neurocomputing"},{"issue":"6","key":"10956_CR23","doi-asserted-by":"publisher","first-page":"2659","DOI":"10.1109\/18.945285","volume":"47","author":"V Kurkova","year":"2001","unstructured":"Kurkova V, Sanguineti M (2001) Bounds on rates of variable-basis and neural-network approximation. IEEE Trans Inf Theory 47(6):2659\u20132665","journal-title":"IEEE Trans Inf Theory"},{"issue":"1","key":"10956_CR24","doi-asserted-by":"publisher","first-page":"549","DOI":"10.1109\/TIT.2010.2090198","volume":"57","author":"G Gnecco","year":"2011","unstructured":"Gnecco G, Sanguineti M (2011) On a variational norm tailored to variable-basis approximation schemes. IEEE Trans Inf Theory 57(1):549\u2013558","journal-title":"IEEE Trans Inf Theory"},{"key":"10956_CR25","doi-asserted-by":"publisher","first-page":"1159","DOI":"10.1007\/s11432-010-0090-0","volume":"6","author":"ZB Xu","year":"2010","unstructured":"Xu ZB, Zhang H, Wang Y, Chang XY, Liang Y (2010) $$L_{1\/2}$$ regularization. Sci China-Inf Sci 6:1159\u20131169","journal-title":"Sci China-Inf Sci"},{"key":"10956_CR26","doi-asserted-by":"publisher","first-page":"359","DOI":"10.1016\/0893-6080(89)90020-8","volume":"2","author":"H Kurt","year":"1989","unstructured":"Kurt H, Maxwell S, Halbert W (1989) Multilayer feedforward networks are universal approximators. Neural Netw 2:359\u2013366","journal-title":"Neural Netw"},{"key":"10956_CR27","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1016\/j.neucom.2018.06.046","volume":"314","author":"F Li","year":"2018","unstructured":"Li F, Zurada JM, Wu W (2018) Smooth Group $$L_{1\/2}$$ regularization for input layer of feedforward neural networks. Neurocomputing 314:109\u2013119","journal-title":"Neurocomputing"},{"key":"10956_CR28","doi-asserted-by":"publisher","first-page":"10979","DOI":"10.1109\/ACCESS.2017.2713389","volume":"5","author":"F Li","year":"2017","unstructured":"Li F, Zurada JM, Liu Y, Wu W (2017) Input layer regularization of multilayer feedforward neural networks. IEEE Access 5:10979\u201310985","journal-title":"IEEE Access"},{"issue":"3","key":"10956_CR29","doi-asserted-by":"publisher","first-page":"1110","DOI":"10.1109\/TNNLS.2020.2980383","volume":"32","author":"J Wang","year":"2021","unstructured":"Wang J, Zhang H, Wang J, Pu YF, Pal NR (2021) Feature selection using a neural network with Group Lasso regularization and controlled redundancy. IEEE Trans Neural Netw Learn Syst 32(3):1110\u20131123","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"10956_CR30","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1016\/j.neucom.2017.02.029","volume":"241","author":"S Scardapane","year":"2017","unstructured":"Scardapane S, Comminiello D, Hussain A, Uncini A (2017) Group sparse regularization for deep neural networks. Neurocomputing 241:81\u201389","journal-title":"Neurocomputing"},{"key":"10956_CR31","doi-asserted-by":"publisher","first-page":"1333","DOI":"10.1109\/TCYB.2019.2950105","volume":"50","author":"XT Xie","year":"2020","unstructured":"Xie XT, Zhang HQ, Wang Chang Q, Wang J, Pal NR (2020) Learning optimized structure of neural networks by hidden node pruning with $$L_1$$ regularization. IEEE T Cybern 50:1333\u20131346","journal-title":"IEEE T Cybern"},{"key":"10956_CR32","doi-asserted-by":"crossref","unstructured":"Formanek A, Hadhzi D (2019) Compressing convolutional neural networks by $$L_0$$ regularization. Proceeding International Conference on Control, Artificial Intelligence, Robotics & Optimization pp 155-162","DOI":"10.1109\/ICCAIRO47923.2019.00032"},{"key":"10956_CR33","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1016\/j.neucom.2017.02.029","volume":"241","author":"S Scardapane","year":"2017","unstructured":"Scardapane S, Comminiello D, Hussain A, Uncini A (2017) Group sparse regularization for deep neural networks. Neurocomputing 241:81\u201389","journal-title":"Neurocomputing"},{"issue":"2","key":"10956_CR34","doi-asserted-by":"publisher","first-page":"383","DOI":"10.1007\/s00521-014-1730-x","volume":"26","author":"HS Zhang","year":"2015","unstructured":"Zhang HS, Tang YL, Liu XD (2015) Batch gradient training method with smoothing regularization for $$L_0$$ feedforward neural networks. Neural Comput & Applic 26(2):383\u2013390","journal-title":"Neural Comput & Applic"},{"key":"10956_CR35","doi-asserted-by":"crossref","unstructured":"Xie Q, Li C, Diao B, An Z, Xu Y (2019) $$L_0$$ regularization based fine-grained neural network pruning method. Proc. Int. Conf. Electron. Comput. Artif. Intell. p 11:1-4","DOI":"10.1109\/ECAI46879.2019.9041962"},{"key":"10956_CR36","doi-asserted-by":"crossref","unstructured":"Wang J, Cai Q, Zurada JM, Chang Q, Zurada JM (2017) Convergence analyses on sparse feedforward neural networks via Group Lasso regularization. Inf Sci 381:250\u2013269","DOI":"10.1016\/j.ins.2016.11.020"},{"key":"10956_CR37","doi-asserted-by":"publisher","first-page":"28742","DOI":"10.1109\/ACCESS.2020.3048235","volume":"9","author":"Q Fan","year":"2021","unstructured":"Fan Q, Peng J, Li H, Lin S (2021) Convergence of a gradient-based learning algorithm with penalty for ridge polynomial neural networks. IEEE Access 9:28742\u201328752","journal-title":"IEEE Access"},{"key":"10956_CR38","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1016\/j.ins.2020.12.014","volume":"553","author":"Q Kang","year":"2021","unstructured":"Kang Q, Fan Q, Zurada JM (2021) Deterministic convergence analysis via smoothing Group Lasso regularization and adaptive momentum for Sigma-Pi-Sigma neural network. Inf Sci 553:66\u201382","journal-title":"Inf Sci"}],"container-title":["Neural Processing Letters"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-022-10956-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11063-022-10956-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-022-10956-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,27]],"date-time":"2023-04-27T02:29:09Z","timestamp":1682562549000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11063-022-10956-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,7,14]]},"references-count":38,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2023,4]]}},"alternative-id":["10956"],"URL":"https:\/\/doi.org\/10.1007\/s11063-022-10956-w","relation":{},"ISSN":["1370-4621","1573-773X"],"issn-type":[{"type":"print","value":"1370-4621"},{"type":"electronic","value":"1573-773X"}],"subject":[],"published":{"date-parts":[[2022,7,14]]},"assertion":[{"value":"1 July 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 July 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}