{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T19:02:54Z","timestamp":1757617374971,"version":"3.44.0"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"7-8","license":[{"start":{"date-parts":[[2025,1,19]],"date-time":"2025-01-19T00:00:00Z","timestamp":1737244800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,19]],"date-time":"2025-01-19T00:00:00Z","timestamp":1737244800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No. 62066001","No. 62066001","No. 62066001"],"award-info":[{"award-number":["No. 62066001","No. 62066001","No. 62066001"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. Mach. Learn. &amp; Cyber."],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s13042-024-02514-8","type":"journal-article","created":{"date-parts":[[2025,1,19]],"date-time":"2025-01-19T02:59:41Z","timestamp":1737255581000},"page":"4367-4387","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A stochastic recursive gradient algorithm integrating momentum and the powerball function with adaptive step sizes"],"prefix":"10.1007","volume":"16","author":[{"given":"Chuandong","family":"Qin","sequence":"first","affiliation":[]},{"given":"Zilin","family":"Cai","sequence":"additional","affiliation":[]},{"given":"Yuhang","family":"Guo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,19]]},"reference":[{"issue":"1","key":"2514_CR1","doi-asserted-by":"publisher","first-page":"305","DOI":"10.1109\/tnnls.2021.3093877","volume":"34","author":"YJ Wang","year":"2023","unstructured":"Wang YJ, Wang T, Yang XB et al (2023) Gradient descent-Barzilai Borwein-based neural network tracking control for nonlinear systems with unknown dynamics. IEEE Trans Neural Netw Learn Syst 34(1):305\u2013315. https:\/\/doi.org\/10.1109\/tnnls.2021.3093877","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"2514_CR2","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.122017","volume":"238","author":"Z Yang","year":"2024","unstructured":"Yang Z, Li X (2024) Powered stochastic optimization with hypergradient descent for large-scale learning systems. Expert Syst Appl 238:122017. https:\/\/doi.org\/10.1016\/j.eswa.2023.122017","journal-title":"Expert Syst Appl"},{"issue":"7","key":"2514_CR3","doi-asserted-by":"publisher","first-page":"1541","DOI":"10.1007\/s13042-019-01055-9","volume":"11","author":"VK Chauhan","year":"2020","unstructured":"Chauhan VK, Sharma A, Dahiya K (2020) Stochastic trust region inexact newton method for large-scale machine learning. Int J Mach Learn Cybern 11(7):1541\u20131555. https:\/\/doi.org\/10.1007\/s13042-019-01055-9","journal-title":"Int J Mach Learn Cybern"},{"issue":"8","key":"2514_CR4","doi-asserted-by":"publisher","first-page":"2237","DOI":"10.1007\/s13042-021-01303-x","volume":"12","author":"L Liu","year":"2021","unstructured":"Liu L, Li P, Chu M et al (2021) Stochastic gradient support vector machine with local structural information for pattern recognition. Int J Mach Learn Cybern 12(8):2237\u20132254. https:\/\/doi.org\/10.1007\/s13042-021-01303-x","journal-title":"Int J Mach Learn Cybern"},{"issue":"2","key":"2514_CR5","doi-asserted-by":"publisher","first-page":"242","DOI":"10.1109\/jstsp.2015.2505682","volume":"10","author":"J Konecny","year":"2016","unstructured":"Konecny J, Liu J, Richt\u00e1rik P et al (2016) Mini-batch semi-stochastic gradient descent in the proximal setting. IEEE J Sel Top Signal Process 10(2):242\u2013255. https:\/\/doi.org\/10.1109\/jstsp.2015.2505682","journal-title":"IEEE J Sel Top Signal Process"},{"key":"2514_CR6","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1016\/j.neucom.2018.06.002","volume":"314","author":"Z Yang","year":"2018","unstructured":"Yang Z, Wang C, Zang Y et al (2018) Mini-batch algorithms with Barzilai\u2013Borwein update step. Neurocomputing 314:177\u2013185. https:\/\/doi.org\/10.1016\/j.neucom.2018.06.002","journal-title":"Neurocomputing"},{"key":"2514_CR7","unstructured":"Pillaud-Vivien L, Rudi A, Bach F (2017) Exponential convergence of testing error for stochastic gradient methods. In: Proceedings of the 31st conference on learning theory, PMLR, pp 250\u2013296"},{"key":"2514_CR8","unstructured":"Rakhlin A, Shamir O, Sridharan K (2012) Making gradient descent optimal for strongly convex stochastic optimization. In: Proceedings of the 29th International Coference on International Conference on Machine Learning, ICML pp 1571\u20131578"},{"key":"2514_CR9","unstructured":"Roux NL, Schmidt M, Bach FJAiNIPS (2012) A stochastic gradient method with an exponential convergence rate for finite training sets. In: Proceedings of the 25th international conference on neural information processing systems, NIPS 2:2663\u20132671"},{"issue":"1","key":"2514_CR10","doi-asserted-by":"publisher","first-page":"83","DOI":"10.1007\/s10107-016-1030-6","volume":"162","author":"M Schmidt","year":"2017","unstructured":"Schmidt M, Le Roux N, Bach F (2017) Minimizing finite sums with the stochastic average gradient. Math Progr 162(1):83\u2013112. https:\/\/doi.org\/10.1007\/s10107-016-1030-6","journal-title":"Math Progr"},{"key":"2514_CR11","unstructured":"Defazio A, Bach F, Lacoste-Julien S (2014) Saga: A fast incremental gradient method with support for non-strongly convex composite objectives. In: Proceedings of the 27th international conference on neural information processing systems, NIPS 1:1646\u20131654"},{"issue":"16","key":"2514_CR12","first-page":"567","volume":"14","author":"S Shalev-Shwartz","year":"2013","unstructured":"Shalev-Shwartz S, Zhang T (2013) Stochastic dual coordinate ascent methods for regularized loss minimization. J Mach Learn Res 14(16):567\u2013599","journal-title":"J Mach Learn Res"},{"key":"2514_CR13","doi-asserted-by":"crossref","unstructured":"Mokhtari A, G\u00fcrb\u00fczbalaban M, Ribeiro A, et\u00a0al (2017) A double incremental aggregated gradient method with linear convergence rate for large-scale optimization. In: 2017 IEEE international conference on acoustics, speech and signal processing, ICASSP pp 4696\u20134700","DOI":"10.1109\/ICASSP.2017.7953047"},{"issue":"4","key":"2514_CR14","doi-asserted-by":"publisher","first-page":"2057","DOI":"10.1137\/140961791","volume":"24","author":"L Xiao","year":"2014","unstructured":"Xiao L, Zhang T (2014) Proximal stochastic gradient method with progressive variance reduction. SIAM J Optim 24(4):2057\u20132075. https:\/\/doi.org\/10.1137\/140961791","journal-title":"SIAM J Optim"},{"issue":"3","key":"2514_CR15","doi-asserted-by":"publisher","first-page":"947","DOI":"10.1007\/s13042-022-01674-9","volume":"14","author":"N Shajoonnezhad","year":"2023","unstructured":"Shajoonnezhad N, Nikanjam A (2023) A stochastic variance-reduced coordinate descent algorithm for learning sparse bayesian network from discrete high-dimensional data. Int J Mach Learn Cybern 14(3):947\u2013958. https:\/\/doi.org\/10.1007\/s13042-022-01674-9","journal-title":"Int J Mach Learn Cybern"},{"key":"2514_CR16","unstructured":"Nguyen LM, Liu J, Scheinberg K, et\u00a0al (2017) Sarah: A novel method for machine learning problems using stochastic recursive gradient. In: Proceedings of the 34th international conference on machine learning, ICML 70:2613\u20132621"},{"key":"2514_CR17","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.117719","volume":"206","author":"Z Yang","year":"2022","unstructured":"Yang Z (2022) Adaptive stochastic conjugate gradient for machine learning. Expert Syst Appl 206:117719. https:\/\/doi.org\/10.1016\/j.eswa.2022.117719","journal-title":"Expert Syst Appl"},{"issue":"23","key":"2514_CR18","doi-asserted-by":"publisher","first-page":"28627","DOI":"10.1007\/s10489-023-05025-1","volume":"53","author":"Z Yang","year":"2023","unstructured":"Yang Z (2023) Stochastic variance reduced gradient with hyper-gradient for non-convex large-scale learning. Appl Intell 53(23):28627\u201328641. https:\/\/doi.org\/10.1007\/s10489-023-05025-1","journal-title":"Appl Intell"},{"issue":"6","key":"2514_CR19","doi-asserted-by":"publisher","first-page":"1598","DOI":"10.1109\/TBDATA.2023.3300546","volume":"9","author":"Z Yang","year":"2023","unstructured":"Yang Z (2023) Adaptive powerball stochastic conjugate gradient for large-scale learning. IEEE Trans Big Data 9(6):1598\u20131606. https:\/\/doi.org\/10.1109\/TBDATA.2023.3300546","journal-title":"IEEE Trans Big Data"},{"key":"2514_CR20","unstructured":"Shamir O, Zhang T (2012) Stochastic gradient descent for non-smooth optimization: Convergence results and optimal averaging schemes. In: Proceedings of the 30th international conference on machine learning, ICML 28:71\u201379"},{"issue":"61","key":"2514_CR21","first-page":"2121","volume":"12","author":"J Duchi","year":"2011","unstructured":"Duchi J, Hazan E, Singer Y (2011) Adaptive subgradient methods for online learning and stochastic optimization. J Mach Learn Res 12(61):2121\u20132159","journal-title":"J Mach Learn Res"},{"issue":"12","key":"2514_CR22","doi-asserted-by":"publisher","first-page":"3785","DOI":"10.1007\/s13042-022-01625-4","volume":"13","author":"R Jie","year":"2022","unstructured":"Jie R, Gao J, Vasnev A et al (2022) Adaptive hierarchical hyper-gradient descent. Int J Mach Learn Cybern 13(12):3785\u20133805. https:\/\/doi.org\/10.1007\/s13042-022-01625-4","journal-title":"Int J Mach Learn Cybern"},{"key":"2514_CR23","doi-asserted-by":"crossref","unstructured":"Zhou B, Liu J, Sun W, et\u00a0al (2021) pbsgd: powered stochastic gradient descent methods for accelerated nonconvex optimization. In: Proceedings of the twenty-ninth international joint conference on artificial intelligence, IJCAI pp 3258\u20133266","DOI":"10.24963\/ijcai.2020\/451"},{"key":"2514_CR24","doi-asserted-by":"crossref","unstructured":"Zhang S, Bailey CP (2022) Accelerated zeroth-order algorithm for stochastic distributed non-convex optimization. In: 2022 American Control Conference, ACC pp 4274\u20134279","DOI":"10.23919\/ACC53348.2022.9867306"},{"issue":"11","key":"2514_CR25","doi-asserted-by":"publisher","first-page":"6993","DOI":"10.1109\/tii.2019.2909328","volume":"16","author":"HT Zhang","year":"2020","unstructured":"Zhang HT, Sun WG, Li YZ et al (2020) A fast optimal power flow algorithm using powerball method. IEEE Trans Ind Inform 16(11):6993\u20137003. https:\/\/doi.org\/10.1109\/tii.2019.2909328","journal-title":"IEEE Trans Ind Inform"},{"key":"2514_CR26","doi-asserted-by":"publisher","first-page":"490","DOI":"10.1016\/j.ins.2021.05.084","volume":"574","author":"ZH Shi","year":"2021","unstructured":"Shi ZH, Wu DR, Guo CF et al (2021) Fcm-rdpa: Tsk fuzzy regression model construction using fuzzy c-means clustering, regularization, droprule, and powerball adabelief. Inf Sci 574:490\u2013504. https:\/\/doi.org\/10.1016\/j.ins.2021.05.084","journal-title":"Inf Sci"},{"issue":"1","key":"2514_CR27","doi-asserted-by":"publisher","first-page":"254","DOI":"10.1109\/tfuzz.2022.3185464","volume":"31","author":"YQ Cui","year":"2023","unstructured":"Cui YQ, Xu YF, Peng RM et al (2023) Layer normalization for tsk fuzzy system optimization in regression problems. IEEE Trans Fuzzy Syst 31(1):254\u2013264. https:\/\/doi.org\/10.1109\/tfuzz.2022.3185464","journal-title":"IEEE Trans Fuzzy Syst"},{"issue":"5","key":"2514_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/0041-5553(64)90137-5","volume":"4","author":"BT Polyak","year":"1964","unstructured":"Polyak BT (1964) Some methods of speeding up the convergence of iteration methods. USSR Comput Math Math Phys 4(5):1\u201317. https:\/\/doi.org\/10.1016\/0041-5553(64)90137-5","journal-title":"USSR Comput Math Math Phys"},{"key":"2514_CR29","volume-title":"Introductory lectures on convex optimization: A basic course","author":"Y Nesterov","year":"2014","unstructured":"Nesterov Y (2014) Introductory lectures on convex optimization: A basic course. Springer Publishing Company, Incorporated"},{"key":"2514_CR30","unstructured":"Ma J, Yarats D (2018) Quasi-hyperbolic momentum and adam for deep learning. ArXiv: 1810.06801"},{"key":"2514_CR31","doi-asserted-by":"publisher","DOI":"10.1007\/s10013-024-00699-7","author":"TH Tran","year":"2024","unstructured":"Tran TH, Tran-Dinh Q, Nguyen LM (2024) Shuffling momentum gradient algorithm for convex optimization. Vietnam J Math. https:\/\/doi.org\/10.1007\/s10013-024-00699-7","journal-title":"Vietnam J Math"},{"key":"2514_CR32","unstructured":"Ruder S (2016) An overview of gradient descent optimization algorithms. ArXiv: 1609.04747"},{"key":"2514_CR33","unstructured":"Shang F, Jiao L, Zhou K, et\u00a0al (2018) Asvrg: Accelerated proximal svrg. In: Proceedings of the 10th asian conference on machine learning, ACML 95:815\u2013830"},{"key":"2514_CR34","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.122295","volume":"238","author":"Z Yang","year":"2024","unstructured":"Yang Z (2024) Sarah-m: A fast stochastic recursive gradient descent algorithm via momentum. Expert Syst Appl 238:122295. https:\/\/doi.org\/10.1016\/j.eswa.2023.122295","journal-title":"Expert Syst Appl"},{"key":"2514_CR35","doi-asserted-by":"publisher","first-page":"218","DOI":"10.1016\/j.ins.2015.03.073","volume":"316","author":"K Sopyla","year":"2015","unstructured":"Sopyla K, Drozda P (2015) Stochastic gradient descent with Barzilai\u2013Borwein update step for svm. Inf Sci 316:218\u2013233. https:\/\/doi.org\/10.1016\/j.ins.2015.03.073","journal-title":"Inf Sci"},{"key":"2514_CR36","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1016\/j.sigpro.2019.02.010","volume":"159","author":"Z Yang","year":"2019","unstructured":"Yang Z, Wang C, Zhang ZM et al (2019) Accelerated stochastic gradient descent with step size selection rules. Signal Process 159:171\u2013186. https:\/\/doi.org\/10.1016\/j.sigpro.2019.02.010","journal-title":"Signal Process"},{"key":"2514_CR37","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1016\/j.engappai.2018.03.017","volume":"72","author":"Z Yang","year":"2018","unstructured":"Yang Z, Wang C, Zhang ZM et al (2018) Random Barzilai\u2013Borwein step size for mini-batch algorithms. Eng Appl Artif Intell 72:124\u2013135. https:\/\/doi.org\/10.1016\/j.engappai.2018.03.017","journal-title":"Eng Appl Artif Intell"},{"key":"2514_CR38","unstructured":"Loizou N, Vaswani S, Laradji I, et\u00a0al (2021) Stochastic polyak step-size for sgd: An adaptive learning rate for fast convergence. In: Proceedings of the 24th international conference on artificial intelligence and statistics, AISTATS 130:1306\u20131314"},{"issue":"1","key":"2514_CR39","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1007\/s10915-022-02084-3","volume":"94","author":"G Franchini","year":"2022","unstructured":"Franchini G, Porta F, Ruggiero V et al (2022) A line search based proximal stochastic gradient algorithm with dynamical variance reduction. J Sci Comput 94(1):23. https:\/\/doi.org\/10.1007\/s10915-022-02084-3","journal-title":"J Sci Comput"},{"issue":"9","key":"2514_CR40","doi-asserted-by":"publisher","first-page":"1433","DOI":"10.1360\/SCM-2019-0795","volume":"51","author":"Liu Yan HCGuo Tiande","year":"2021","unstructured":"Liu Yan HCGuo Tiande (2021) Analysis and improvement for a class of variance reduced methods. SCI SIN Math 51(9):1433\u20131450. https:\/\/doi.org\/10.1360\/SCM-2019-0795","journal-title":"SCI SIN Math"},{"key":"2514_CR41","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2020.114336","volume":"169","author":"Z Yang","year":"2021","unstructured":"Yang Z (2021) On the step size selection in variance-reduced algorithm for nonconvex optimization. Expert Syst Appl 169:114336. https:\/\/doi.org\/10.1016\/j.eswa.2020.114336","journal-title":"Expert Syst Appl"},{"key":"2514_CR42","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2023.104230","volume":"143","author":"CY He","year":"2023","unstructured":"He CY, Zhang YT, Zhu DY et al (2023) A mini-batch algorithm for large-scale learning problems with adaptive step size. Digit Signal Process 143:104230. https:\/\/doi.org\/10.1016\/j.dsp.2023.104230","journal-title":"Digit Signal Process"},{"key":"2514_CR43","doi-asserted-by":"publisher","unstructured":"Yang Z, Ma L (2021) Adaptive stochastic gradient descent for large-scale learning problems https:\/\/doi.org\/10.21203\/rs.3.rs-1066512\/v1","DOI":"10.21203\/rs.3.rs-1066512\/v1"},{"issue":"241","key":"2514_CR44","first-page":"1","volume":"24","author":"Z Yang","year":"2023","unstructured":"Yang Z (2023) Improved powered stochastic optimization algorithms for large-scale machine learning. J Mach Learn Res 24(241):1\u201329","journal-title":"J Mach Learn Res"},{"issue":"3","key":"2514_CR45","doi-asserted-by":"publisher","first-page":"561","DOI":"10.1007\/s10589-022-00375-x","volume":"82","author":"LM Nguyen","year":"2022","unstructured":"Nguyen LM, van Dijk M, Phan DT et al (2022) Finite-sum smooth optimization with sarah. Comput Optim Appl 82(3):561\u2013593. https:\/\/doi.org\/10.1007\/s10589-022-00375-x","journal-title":"Comput Optim Appl"},{"key":"2514_CR46","unstructured":"Nitanda A (2014) Stochastic proximal gradient descent with acceleration techniques. In: Proceedings of the 27th international conference on neural information processing systems, NIPS 1:1574\u20131582"},{"issue":"1","key":"2514_CR47","doi-asserted-by":"publisher","first-page":"1","DOI":"10.2140\/pjm.1966.16.1","volume":"16","author":"L Armijo","year":"1966","unstructured":"Armijo L (1966) Minimization of functions having lipschitz continuous first partial derivatives. Pac J Math 16(1):1\u20133","journal-title":"Pac J Math"},{"key":"2514_CR48","doi-asserted-by":"publisher","first-page":"367","DOI":"10.1016\/S0377-0427(99)00276-9","volume":"114","author":"M Vrahatis","year":"2000","unstructured":"Vrahatis M, Androulakis G, Lambrinos JN et al (2000) A class of gradient unconstrained minimization algorithms with adaptive stepsize. J Comput Appl Math 114:367\u2013386. https:\/\/doi.org\/10.1016\/S0377-0427(99)00276-9","journal-title":"J Comput Appl Math"},{"issue":"2","key":"2514_CR49","doi-asserted-by":"publisher","first-page":"634","DOI":"10.1137\/21M1394308","volume":"4","author":"S Horv\u00e1th","year":"2022","unstructured":"Horv\u00e1th S, Lei L, Richt\u00e1rik P et al (2022) Adaptivity of stochastic gradient methods for nonconvex optimization. SIAM J Math Data Sci 4(2):634\u2013648. https:\/\/doi.org\/10.1137\/21M1394308","journal-title":"SIAM J Math Data Sci"},{"key":"2514_CR50","unstructured":"Tan C, Ma S, Dai YH, et\u00a0al (2016) Barzilai-borwein step size for stochastic gradient descent. In: Proceedings of the 30th international conference on neural information processing systems, NIPS pp 685\u2013693"},{"key":"2514_CR51","unstructured":"Li B, Wang L, Giannakis GB (2020) Almost tune-free variance reduction. In: Proceedings of the 37th international conference on machine learning, ICML 119:5969\u20135978"},{"key":"2514_CR52","unstructured":"Kingma DP, Ba J (2014) Adam: A method for stochastic optimization. arXiv e-prints p arXiv:1412.6980"}],"container-title":["International Journal of Machine Learning and Cybernetics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-024-02514-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13042-024-02514-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-024-02514-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T04:13:44Z","timestamp":1757132024000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13042-024-02514-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,19]]},"references-count":52,"journal-issue":{"issue":"7-8","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["2514"],"URL":"https:\/\/doi.org\/10.1007\/s13042-024-02514-8","relation":{},"ISSN":["1868-8071","1868-808X"],"issn-type":[{"type":"print","value":"1868-8071"},{"type":"electronic","value":"1868-808X"}],"subject":[],"published":{"date-parts":[[2025,1,19]]},"assertion":[{"value":"22 May 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 December 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 January 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there are no conflict of interest regarding the publication of this article, and they have no potential financial interests that could influence the reporting of the competing work described herein.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"We consent to the publication of the paper.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"Materials will be made available on request.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Materials availability"}},{"value":"Code will be made available on request.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Code availability"}}]}}