{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T23:37:13Z","timestamp":1772321833129,"version":"3.50.1"},"reference-count":32,"publisher":"Springer Science and Business Media LLC","issue":"7-8","license":[{"start":{"date-parts":[[2025,1,9]],"date-time":"2025-01-09T00:00:00Z","timestamp":1736380800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,9]],"date-time":"2025-01-09T00:00:00Z","timestamp":1736380800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Guangxi Science and Technology base and Talent Project","award":["(Grant No. AD22080047)"],"award-info":[{"award-number":["(Grant No. AD22080047)"]}]},{"name":"Guangxi Science and Technology base and Talent Project","award":["(Grant No. AD22080047)"],"award-info":[{"award-number":["(Grant No. AD22080047)"]}]},{"name":"Guangxi Science and Technology base and Talent Project","award":["(Grant No. AD22080047)"],"award-info":[{"award-number":["(Grant No. AD22080047)"]}]},{"name":"the National Natural Science Foundation of Guangxi Province","award":["(Grant No. 2023GXNFSBA 026063)"],"award-info":[{"award-number":["(Grant No. 2023GXNFSBA 026063)"]}]},{"name":"the National Natural Science Foundation of Guangxi Province","award":["(Grant No. 2023GXNFSBA 026063)"],"award-info":[{"award-number":["(Grant No. 2023GXNFSBA 026063)"]}]},{"name":"the National Natural Science Foundation of Guangxi Province","award":["(Grant No. 2023GXNFSBA 026063)"],"award-info":[{"award-number":["(Grant No. 2023GXNFSBA 026063)"]}]},{"name":"the special foundation for Guangxi Ba Gui Scholars"},{"name":"2024 Graduate Innovative Programs Establishment","award":["(Grant No. ZX01030031124006)"],"award-info":[{"award-number":["(Grant No. ZX01030031124006)"]}]},{"name":"the major talent project of Guangxi","award":["(Grant No. GXR-6BG242404)"],"award-info":[{"award-number":["(Grant No. GXR-6BG242404)"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. Mach. Learn. &amp; Cyber."],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s13042-024-02524-6","type":"journal-article","created":{"date-parts":[[2025,1,9]],"date-time":"2025-01-09T08:32:16Z","timestamp":1736411536000},"page":"4545-4559","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["A stochastic recursive gradient algorithm with inertial extrapolation for non-convex problems and machine learning"],"prefix":"10.1007","volume":"16","author":[{"given":"Zhongyu","family":"Mo","sequence":"first","affiliation":[]},{"given":"Chen","family":"Ouyang","sequence":"additional","affiliation":[]},{"given":"Hongtruong","family":"Pham","sequence":"additional","affiliation":[]},{"given":"Gonglin","family":"Yuan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,9]]},"reference":[{"key":"2524_CR1","doi-asserted-by":"crossref","unstructured":"Robbins H, Monro S (1951) A stochastic approximation method. The Annals of Mathematical Statistics, pp 00\u2013407","DOI":"10.1214\/aoms\/1177729586"},{"key":"2524_CR2","unstructured":"Nguyen L, Liu J, Scheinberg K, Tak\u00e1\u010d M (2017) SARAH: a novel method for machine learning problems using stochastic recursive gradient. In: International Conference on Machine Learning. PMLR, pp 2613\u20132621"},{"issue":"1","key":"2524_CR3","doi-asserted-by":"publisher","first-page":"237","DOI":"10.1080\/10556788.2020.1818081","volume":"36","author":"L Nguyen","year":"2021","unstructured":"Nguyen L, Scheinberg K, Tak\u00e1\u010d M (2021) Inexact SARAH algorithm for stochastic optimization. Optim Methods Softw 36(1):237\u2013258","journal-title":"Optim Methods Softw"},{"key":"2524_CR4","unstructured":"Nguyen L, Liu J, Scheinberg K, Tak\u00e1\u010d M (2017) Stochastic recursive gradient algorithm for nonconvex optimization. arXiv preprint arXiv:1705.07261"},{"issue":"3","key":"2524_CR5","doi-asserted-by":"publisher","first-page":"107","DOI":"10.1007\/s11222-024-10409-5","volume":"34","author":"C Ouyang","year":"2024","unstructured":"Ouyang C, Lu C, Zhao X, Huang R, Yuan G, Jiang Y (2024) Stochastic three-term conjugate gradient method with variance technique for non-convex learning. Stat Comput 34(3):107","journal-title":"Stat Comput"},{"key":"2524_CR6","unstructured":"Roux N, Schmidt M, Bach F (2012) A stochastic gradient method with an exponential convergence rate for finite training sets. In: Advances in neural information processing systems, vol 25"},{"key":"2524_CR7","unstructured":"Defazio A, Bach F, Lacoste-Julien S (2014) SAGA: a fast incremental gradient method with support for non-strongly convex composite objectives. In: Advances in neural information processing systems, vol 27"},{"key":"2524_CR8","unstructured":"Johnson R, Zhang T (2013) Accelerating stochastic gradient descent using predictive variance reduction. In: Advances in neural information processing systems, vol 26"},{"key":"2524_CR9","doi-asserted-by":"publisher","first-page":"189","DOI":"10.1186\/s13660-021-02719-3","volume":"1","author":"A Ibrahim","year":"2021","unstructured":"Ibrahim A, Kumam P, Abubakar A, Abubakar J (2021) A method with inertial extrapolation step for convex constrained monotone equations. J Inequal Appl 1:189","journal-title":"J Inequal Appl"},{"issue":"2","key":"2524_CR10","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1007\/s10915-024-02518-0","volume":"99","author":"J Bai","year":"2024","unstructured":"Bai J, Jia L, Peng Z (2024) A new insight on augmented Lagrangian method with applications in machine learning. J Sci Comput 99(2):53","journal-title":"J Sci Comput"},{"issue":"2","key":"2524_CR11","doi-asserted-by":"publisher","first-page":"479","DOI":"10.1007\/s10589-021-00338-8","volume":"81","author":"J Bai","year":"2022","unstructured":"Bai J, Hager W, Zhang H (2022) An inexact accelerated stochastic ADMM for separable convex optimization. Comput Optim Appl 81(2):479\u2013518","journal-title":"Comput Optim Appl"},{"key":"2524_CR12","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.121556","volume":"238","author":"R Huang","year":"2024","unstructured":"Huang R, Qin Y, Liu K, Yuan G (2024) Biased stochastic conjugate gradient algorithm with adaptive step size for nonconvex problems. Expert Syst Appl 238:121556","journal-title":"Expert Syst Appl"},{"issue":"2","key":"2524_CR13","doi-asserted-by":"publisher","DOI":"10.1088\/0266-5611\/29\/2\/025011","volume":"29","author":"P Chen","year":"2013","unstructured":"Chen P, Huang J, Zhang X (2013) A primal-dual fixed point algorithm for convex separable minimization with applications to image restoration. Inverse Probl 29(2):025011","journal-title":"Inverse Probl"},{"issue":"3","key":"2524_CR14","doi-asserted-by":"publisher","first-page":"862","DOI":"10.1137\/110849456","volume":"22","author":"H Iiduka","year":"2012","unstructured":"Iiduka H (2012) Iterative algorithm for triple-hierarchical constrained nonconvex optimization problem and its application to network bandwidth allocation. SIAM J Optim 22(3):862\u2013878","journal-title":"SIAM J Optim"},{"issue":"2","key":"2524_CR15","doi-asserted-by":"publisher","first-page":"387","DOI":"10.1080\/02331934.2020.1716752","volume":"70","author":"L Jolaoso","year":"2021","unstructured":"Jolaoso L, Alakoya T, Taiwo A, Mewomo O (2021) Inertial extragradient method via viscosity approximation approach for solving equilibrium problem in Hilbert space. Optimization 70(2):387\u2013412","journal-title":"Optimization"},{"issue":"5","key":"2524_CR16","doi-asserted-by":"publisher","first-page":"818","DOI":"10.3390\/math8050818","volume":"8","author":"J Abubakar","year":"2020","unstructured":"Abubakar J, Kumam P, Hassan Ibrahim A, Padcharoen A (2020) Relaxed inertial Tseng\u2019s type method for solving the inclusion problem with application to image restoration. Mathematics 8(5):818","journal-title":"Mathematics"},{"issue":"4","key":"2524_CR17","doi-asserted-by":"publisher","first-page":"609","DOI":"10.3390\/math8040609","volume":"8","author":"J Abubakar","year":"2020","unstructured":"Abubakar J, Kumam P, Rehman H, Hassan Ibrahim A (2020) Inertial iterative schemes with variable step sizes for variational inequality problem involving pseudomonotone operator. Mathematics 8(4):609","journal-title":"Mathematics"},{"issue":"1","key":"2524_CR18","first-page":"166","volume":"18","author":"J Abubakar","year":"2019","unstructured":"Abubakar J, Sombut K, Ibrahim A (2019) An accelerated subgradient extragradient algorithm for strongly pseudomonotone variational inequality problems. Thai J Math 18(1):166\u2013187","journal-title":"Thai J Math"},{"issue":"7","key":"2524_CR19","doi-asserted-by":"publisher","first-page":"309","DOI":"10.1007\/s40314-022-02019-6","volume":"41","author":"J Jian","year":"2022","unstructured":"Jian J, Yin J, Tang C, Han D (2022) A family of inertial derivative-free projection methods for constrained nonlinear pseudo-monotone equations with applications. Comput Appl Math 41(7):309","journal-title":"Comput Appl Math"},{"key":"2524_CR20","doi-asserted-by":"publisher","first-page":"129","DOI":"10.1007\/s10589-019-00073-1","volume":"73","author":"Z Wu","year":"2019","unstructured":"Wu Z, Li M (2019) General inertial proximal gradient method for a class of nonconvex nonsmooth optimization problems. Comput Optim Appl 73:129\u2013158","journal-title":"Comput Optim Appl"},{"issue":"4","key":"2524_CR21","doi-asserted-by":"publisher","first-page":"2057","DOI":"10.1137\/140961791","volume":"24","author":"L Xiao","year":"2014","unstructured":"Xiao L, Zhang T (2014) A proximal stochastic gradient method with progressive variance reduction. SIAM J Optim 24(4):2057\u20132075","journal-title":"SIAM J Optim"},{"key":"2524_CR22","doi-asserted-by":"publisher","first-page":"26","DOI":"10.1186\/s13660-023-02922-4","volume":"1","author":"S Sun","year":"2023","unstructured":"Sun S, He L (2023) General inertial proximal stochastic variance reduction gradient for nonconvex nonsmooth optimization. J Inequal Appl 1:26","journal-title":"J Inequal Appl"},{"key":"2524_CR23","doi-asserted-by":"publisher","first-page":"245","DOI":"10.1007\/s11075-018-0603-2","volume":"82","author":"J Liu","year":"2019","unstructured":"Liu J, Feng Y (2019) A derivative-free iterative method for nonlinear monotone equations with convex constraints. Numer Algorithms 82:245\u2013262","journal-title":"Numer Algorithms"},{"issue":"5","key":"2524_CR24","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/0041-5553(64)90137-5","volume":"4","author":"B Polyak","year":"1964","unstructured":"Polyak B (1964) Some methods of speeding up the convergence of iteration methods. USSR Comput Math Math Phys 4(5):1\u201317","journal-title":"USSR Comput Math Math Phys"},{"issue":"3","key":"2524_CR25","doi-asserted-by":"publisher","first-page":"773","DOI":"10.1137\/S1052623403427859","volume":"14","author":"F Alvarez","year":"2004","unstructured":"Alvarez F (2004) Weak convergence of a relaxed and inertial hybrid projection-proximal point algorithm for maximal monotone operators in Hilbert space. SIAM J Optim 14(3):773\u2013782","journal-title":"SIAM J Optim"},{"key":"2524_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.117719","volume":"206","author":"Z Yang","year":"2022","unstructured":"Yang Z (2022) Adaptive stochastic conjugate gradient for machine learning. Expert Syst Appl 206:117719","journal-title":"Expert Syst Appl"},{"key":"2524_CR27","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.122295","volume":"238","author":"Z Yang","year":"2024","unstructured":"Yang Z (2024) SARAH-M: a fast stochastic recursive gradient descent algorithm via momentum. Expert Syst Appl 238:122295","journal-title":"Expert Syst Appl"},{"issue":"3","key":"2524_CR28","doi-asserted-by":"publisher","first-page":"1727","DOI":"10.1007\/s11063-021-10705-5","volume":"54","author":"C Castera","year":"2022","unstructured":"Castera C, Bolte J, F\u00e9votte C, Pauwels E (2022) Second-order step-size tuning of SGD for non-convex optimization. Neural Process Lett 54(3):1727\u20131752","journal-title":"Neural Process Lett"},{"issue":"4","key":"2524_CR29","doi-asserted-by":"publisher","first-page":"864","DOI":"10.1016\/0041-5553(63)90382-3","volume":"3","author":"B Polyak","year":"1963","unstructured":"Polyak B (1963) Gradient methods for the minimisation of functionals. USSR Comput Math Math Phys 3(4):864\u2013878","journal-title":"USSR Comput Math Math Phys"},{"issue":"1","key":"2524_CR30","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1007\/s10107-006-0706-8","volume":"108","author":"Y Nesterov","year":"2006","unstructured":"Nesterov Y, Polyak B (2006) Cubic regularization of Newton method and its global performance. Math Program 108(1):177\u2013205","journal-title":"Math Program"},{"key":"2524_CR31","unstructured":"Reddi S, Hefny A, Sra S, Poczos B, Smola A (2016) Stochastic variance reduction for nonconvex optimization. In: International conference on machine learning. PMLR, pp 314\u2013323"},{"issue":"110","key":"2524_CR32","first-page":"1","volume":"21","author":"N Pham","year":"2020","unstructured":"Pham N, Nguyen L, Phan D, Tran-Dinh Q (2020) ProxSARAH: an efficient algorithmic framework for stochastic composite nonconvex optimization. J Mach Learn Res 21(110):1\u201348","journal-title":"J Mach Learn Res"}],"container-title":["International Journal of Machine Learning and Cybernetics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-024-02524-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13042-024-02524-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-024-02524-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T03:30:48Z","timestamp":1757129448000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13042-024-02524-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,9]]},"references-count":32,"journal-issue":{"issue":"7-8","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["2524"],"URL":"https:\/\/doi.org\/10.1007\/s13042-024-02524-6","relation":{},"ISSN":["1868-8071","1868-808X"],"issn-type":[{"value":"1868-8071","type":"print"},{"value":"1868-808X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1,9]]},"assertion":[{"value":"23 May 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 December 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 January 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}