{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T05:47:23Z","timestamp":1751348843773,"version":"3.37.3"},"reference-count":31,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2020,3,7]],"date-time":"2020-03-07T00:00:00Z","timestamp":1583539200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2020,3,7]],"date-time":"2020-03-07T00:00:00Z","timestamp":1583539200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CCF-1717391"],"award-info":[{"award-number":["CCF-1717391"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Comput Optim Appl"],"published-print":{"date-parts":[[2020,6]]},"DOI":"10.1007\/s10589-020-00183-1","type":"journal-article","created":{"date-parts":[[2020,3,7]],"date-time":"2020-03-07T18:02:46Z","timestamp":1583604166000},"page":"347-380","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["Accelerating incremental gradient optimization with curvature information"],"prefix":"10.1007","volume":"76","author":[{"given":"Hoi-To","family":"Wai","sequence":"first","affiliation":[]},{"given":"Wei","family":"Shi","sequence":"additional","affiliation":[]},{"given":"C\u00e9sar A.","family":"Uribe","sequence":"additional","affiliation":[]},{"given":"Angelia","family":"Nedi\u0107","sequence":"additional","affiliation":[]},{"given":"Anna","family":"Scaglione","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,3,7]]},"reference":[{"key":"183_CR1","unstructured":"Agarwal, A., Bottou, L.: A lower bound for the optimization of finite sums. In: International Conference on Machine Learning, pp. 78\u201386 (2015)"},{"key":"183_CR2","unstructured":"Arjevani, Y., Shamir, O.: Dimension-free iteration complexity of finite sum optimization problems. In: Advances in Neural Information\nProcessing Systems 29, pp. 3540\u20133548 (2016)"},{"key":"183_CR3","volume-title":"Nonlinear Programming","author":"DP Bertsekas","year":"1999","unstructured":"Bertsekas, D.P.: Nonlinear Programming. Athena Scientific, Belmont (1999)"},{"issue":"1\u201338","key":"183_CR4","first-page":"3","volume":"2010","author":"DP Bertsekas","year":"2011","unstructured":"Bertsekas, D.P.: Incremental gradient, subgradient, and proximal methods for convex optimization: a survey. Optim. Mach. Learn. 2010(1\u201338), 3 (2011)","journal-title":"Optim. Mach. Learn."},{"issue":"1","key":"183_CR5","doi-asserted-by":"publisher","first-page":"29","DOI":"10.1137\/040615961","volume":"18","author":"D Blatt","year":"2007","unstructured":"Blatt, D., Hero, A.O., Gauchman, H.: A convergent incremental gradient method with a constant step size. SIAM J. Optim. 18(1), 29\u201351 (2007)","journal-title":"SIAM J. Optim."},{"issue":"2","key":"183_CR6","doi-asserted-by":"publisher","first-page":"223","DOI":"10.1137\/16M1080173","volume":"60","author":"L Bottou","year":"2018","unstructured":"Bottou, L., Curtis, F.E., Nocedal, J.: Optimization methods for large-scale machine learning. SIAM Rev. 60(2), 223\u2013311 (2018)","journal-title":"SIAM Rev."},{"issue":"3\u20134","key":"183_CR7","doi-asserted-by":"publisher","first-page":"231","DOI":"10.1561\/2200000050","volume":"8","author":"S Bubeck","year":"2015","unstructured":"Bubeck, S., et al.: Convex optimization: algorithms and complexity. Found. Trends Mach. Learn. 8(3\u20134), 231\u2013357 (2015)","journal-title":"Found. Trends Mach. Learn."},{"key":"183_CR8","doi-asserted-by":"publisher","first-page":"27:1","DOI":"10.1145\/1961189.1961199","volume":"2","author":"CC Chang","year":"2011","unstructured":"Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines. ACM Trans. Intell. Syst. Technol. 2, 27:1\u201327:27 (2011)","journal-title":"ACM Trans. Intell. Syst. Technol."},{"key":"183_CR9","unstructured":"Defazio, A., Bach, F., Lacoste-Julien, S.: Saga: a fast incremental gradient method with support for non-strongly convex composite objectives. In: Advances in Neural Information Processing Systems 27, pp. 1646\u20131654 (2014)"},{"key":"183_CR10","doi-asserted-by":"crossref","unstructured":"Feyzmahdavian, H.R., Aytekin, A., Johansson, M.: A delayed proximal gradient method with linear convergence rate. In: IEEE International Workshop on Machine Learning for Signal Processing (MLSP), pp. 1\u20136. IEEE (2014)","DOI":"10.1109\/MLSP.2014.6958872"},{"key":"183_CR11","unstructured":"Gower, R.M., Roux, N.L., Bach, F.: Tracking the gradients using the hessian: a new look at variance reducing stochastic methods. In: AISTATS (2018)"},{"issue":"1","key":"183_CR12","doi-asserted-by":"publisher","first-page":"283","DOI":"10.1007\/s10107-015-0897-y","volume":"151","author":"M G\u00fcrb\u00fczbalaban","year":"2015","unstructured":"G\u00fcrb\u00fczbalaban, M., Ozdaglar, A., Parrilo, P.: A globally convergent incremental newton method. Math. Program. 151(1), 283\u2013313 (2015)","journal-title":"Math. Program."},{"key":"183_CR13","doi-asserted-by":"publisher","DOI":"10.1007\/s10107-019-01440-w","author":"M G\u00fcrb\u00fczbalaban","year":"2019","unstructured":"G\u00fcrb\u00fczbalaban, M., Ozdaglar, A., Parrilo, P.: Why random reshuffling beats stochastic gradient descent. Math. Program. https:\/\/doi.org\/10.1007\/s10107-019-01440-w (2019)","journal-title":"Math. Program."},{"issue":"2","key":"183_CR14","doi-asserted-by":"publisher","first-page":"1035","DOI":"10.1137\/15M1049695","volume":"27","author":"M G\u00fcrb\u00fczbalaban","year":"2017","unstructured":"G\u00fcrb\u00fczbalaban, M., Ozdaglar, A., Parrilo, P.: On the convergence rate of incremental aggregated gradient algorithms. SIAM J. Optim. 27(2), 1035\u20131048 (2017)","journal-title":"SIAM J. Optim."},{"key":"183_CR15","doi-asserted-by":"publisher","first-page":"167","DOI":"10.1007\/s10107-017-1173-0","volume":"171","author":"G Lan","year":"2018","unstructured":"Lan, G., Zhou, Y.: An optimal randomized incremental gradient method. Math. Program. 171, 167\u2013215 (2018)","journal-title":"Math. Program."},{"issue":"2","key":"183_CR16","doi-asserted-by":"publisher","first-page":"829","DOI":"10.1137\/140957639","volume":"25","author":"J Mairal","year":"2015","unstructured":"Mairal, J.: Incremental majorization-minimization optimization with application to large-scale machine learning. SIAM J. Optim. 25(2), 829\u2013855 (2015)","journal-title":"SIAM J. Optim."},{"issue":"2","key":"183_CR17","doi-asserted-by":"publisher","first-page":"1670","DOI":"10.1137\/17M1122943","volume":"28","author":"A Mokhtari","year":"2018","unstructured":"Mokhtari, A., Eisen, M., Ribeiro, A.: Iqn: an incremental quasi-newton method with local superlinear convergence rate. SIAM J. Optim. 28(2), 1670\u20131698 (2018)","journal-title":"SIAM J. Optim."},{"key":"183_CR18","series-title":"Applied Optimization","volume-title":"Stochastic Optimization: Algorithms and Applications","author":"A Nedi\u0107","year":"2001","unstructured":"Nedi\u0107 A., Bertsekas D.: Convergence Rate of Incremental Subgradient Algorithms. In: Uryasev S., Pardalos P.M. (eds) Stochastic Optimization: Algorithms and Applications. Applied Optimization, vol. 54. Springer, Boston, MA (2001)"},{"issue":"1","key":"183_CR19","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1137\/S1052623499362111","volume":"12","author":"A Nedic","year":"2001","unstructured":"Nedic, A., Bertsekas, D.P.: Incremental subgradient methods for nondifferentiable optimization. SIAM J. Optim. 12(1), 109\u2013138 (2001)","journal-title":"SIAM J. Optim."},{"key":"183_CR20","volume-title":"Introductory Lectures on Convex Optimization: A Basic Course","author":"Y Nesterov","year":"2013","unstructured":"Nesterov, Y.: Introductory Lectures on Convex Optimization: A Basic Course, vol. 87. Springer, Berlin (2013)"},{"key":"183_CR21","unstructured":"Nitanda, A.: Stochastic proximal gradient descent with acceleration techniques. In: Advances in Neural Information Processing Systems, pp. 1574\u20131582 (2014)"},{"issue":"3","key":"183_CR22","doi-asserted-by":"publisher","first-page":"400","DOI":"10.1214\/aoms\/1177729586","volume":"22","author":"H Robbins","year":"1951","unstructured":"Robbins, H., Monro, S.: A stochastic approximation method. Ann. Math. Stat. 22(3), 400\u2013407 (1951)","journal-title":"Ann. Math. Stat."},{"key":"183_CR23","unstructured":"Rodomanov, A., Kropotov, D.: A superlinearly-convergent proximal Newton-type method for the optimization of finite sums. In: International Conference on Machine Learning, pp. 2597\u20132605 (2016)"},{"issue":"1\u20132","key":"183_CR24","doi-asserted-by":"publisher","first-page":"83","DOI":"10.1007\/s10107-016-1030-6","volume":"162","author":"M Schmidt","year":"2017","unstructured":"Schmidt, M., Le Roux, N., Bach, F.: Minimizing finite sums with the stochastic average gradient. Math. Program. 162(1\u20132), 83\u2013112 (2017)","journal-title":"Math. Program."},{"key":"183_CR25","unstructured":"Schmidt, M., Roux, N.L., Bach, F.R.: Convergence rates of inexact proximal-gradient methods for convex optimization. In: Advances in Neural Information Processing Systems 24, pp. 1458\u20131466 (2011)"},{"issue":"4","key":"183_CR26","doi-asserted-by":"publisher","first-page":"963","DOI":"10.1080\/10556788.2017.1296439","volume":"32","author":"AMC So","year":"2017","unstructured":"So, A.M.C., Zhou, Z.: Non-asymptotic convergence analysis of inexact gradient methods for machine learning without strong convexity. Optim. Methods Softw. 32(4), 963\u2013992 (2017)","journal-title":"Optim. Methods Softw."},{"key":"183_CR27","doi-asserted-by":"crossref","unstructured":"Vanli, N.D., G\u00fcrb\u00fczbalaban, M., Ozdaglar, A.: A stronger convergence result on the proximal incremental aggregated gradient method. arXiv preprint arXiv:1611.08022 (2016)","DOI":"10.1109\/CDC.2016.7798265"},{"issue":"5","key":"183_CR28","doi-asserted-by":"publisher","first-page":"988","DOI":"10.1109\/72.788640","volume":"10","author":"VN Vapnik","year":"1999","unstructured":"Vapnik, V.N.: An overview of statistical learning theory. IEEE Trans. Neural Netw. 10(5), 988\u2013999 (1999)","journal-title":"IEEE Trans. Neural Netw."},{"key":"183_CR29","doi-asserted-by":"crossref","unstructured":"Wai, H.T., Shi, W., Nedi\u0107, A., Scaglione, A.: Curvature-aided incremental aggregated gradient method. In: Proceedings of Allerton (2017)","DOI":"10.1109\/ALLERTON.2017.8262782"},{"issue":"4","key":"183_CR30","doi-asserted-by":"publisher","first-page":"2057","DOI":"10.1137\/140961791","volume":"24","author":"L Xiao","year":"2014","unstructured":"Xiao, L., Zhang, T.: A proximal stochastic gradient method with progressive variance reduction. SIAM J. Optim. 24(4), 2057\u20132075 (2014)","journal-title":"SIAM J. Optim."},{"key":"183_CR31","unstructured":"Zheng, S., Meng, Q., Wang, T., Chen, W., Yu, N., Ma, Z.M., Liu, T.Y.: Asynchronous stochastic gradient descent with delay compensation. In: Proceedings of the 34th International Conference on Machine Learning, vol. 70, pp. 4120\u20134129. JMLR.org (2017)"}],"container-title":["Computational Optimization and Applications"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s10589-020-00183-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s10589-020-00183-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s10589-020-00183-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,3,7]],"date-time":"2021-03-07T00:45:32Z","timestamp":1615077932000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s10589-020-00183-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,3,7]]},"references-count":31,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2020,6]]}},"alternative-id":["183"],"URL":"https:\/\/doi.org\/10.1007\/s10589-020-00183-1","relation":{},"ISSN":["0926-6003","1573-2894"],"issn-type":[{"type":"print","value":"0926-6003"},{"type":"electronic","value":"1573-2894"}],"subject":[],"published":{"date-parts":[[2020,3,7]]},"assertion":[{"value":"22 June 2019","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 March 2020","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}