{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T23:57:52Z","timestamp":1769817472177,"version":"3.49.0"},"reference-count":19,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2014,9,16]],"date-time":"2014-09-16T00:00:00Z","timestamp":1410825600000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Optim Lett"],"published-print":{"date-parts":[[2015,6]]},"DOI":"10.1007\/s11590-014-0795-x","type":"journal-article","created":{"date-parts":[[2014,9,15]],"date-time":"2014-09-15T17:55:37Z","timestamp":1410803737000},"page":"961-979","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":33,"title":["Restricted strong convexity and its applications to convergence analysis of gradient-type methods in convex optimization"],"prefix":"10.1007","volume":"9","author":[{"given":"Hui","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Lizhi","family":"Cheng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2014,9,16]]},"reference":[{"key":"795_CR1","doi-asserted-by":"crossref","first-page":"183","DOI":"10.1137\/080716542","volume":"2","author":"A Beck","year":"2009","unstructured":"Beck, A., Teboulle, M.: A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2, 183\u2013202 (2009)","journal-title":"SIAM J. Imaging Sci."},{"key":"795_CR2","volume-title":"Data Networks","author":"DP Bertsekas","year":"1987","unstructured":"Bertsekas, D.P., Gallager, R.: Data Networks. Prentice-Hall, Englewood Cliffs (1987)"},{"issue":"1","key":"795_CR3","doi-asserted-by":"crossref","first-page":"29","DOI":"10.1137\/040615961","volume":"18","author":"D Blatt","year":"2007","unstructured":"Blatt, D., Hero, A.O., Gauchman, H.: A convergent incremental gradient method with a constant step size. SIAM J. Optim. 18(1), 29\u201351 (2007)","journal-title":"SIAM J. Optim."},{"issue":"1","key":"795_CR4","doi-asserted-by":"crossref","first-page":"127","DOI":"10.1007\/s10107-012-0572-5","volume":"134","author":"RH Byrd","year":"2012","unstructured":"Byrd, R.H., Chin, G.M., Nocedal, J., Wu, Y.: Sample size selection in optimization models for machine learning. Math. Progr. Ser. B 134(1), 127\u2013155 (2012)","journal-title":"Math. Progr. Ser. B"},{"issue":"3","key":"795_CR5","doi-asserted-by":"crossref","first-page":"1380","DOI":"10.1137\/110830629","volume":"34","author":"MP Friedlander","year":"2012","unstructured":"Friedlander, M.P., Schmidt, M.: Hybrid deterministic-stochastic methods for data fitting. SIAM J. Sci. Comput. 34(3), 1380\u20131405 (2012)","journal-title":"SIAM J. Sci. Comput."},{"issue":"2","key":"795_CR6","doi-asserted-by":"crossref","first-page":"1059","DOI":"10.1137\/120863290","volume":"6","author":"MJ Lai","year":"2013","unstructured":"Lai, M.J., Yin, W.: Augmented $$\\ell _1$$ \u2113 1 and nuclear-norm models with a globally linearly convergent algorithm. SIAM J. Imaging Sci. 6(2), 1059\u20131091 (2013)","journal-title":"SIAM J. Imaging Sci."},{"issue":"1","key":"795_CR7","doi-asserted-by":"crossref","first-page":"155","DOI":"10.1137\/0803008","volume":"3","author":"W Li","year":"1993","unstructured":"Li, W.: Remarks on convergence of the matrix splitting algorithm for the symmetric linear complementarity problem. SIAM J. Optim. 3(1), 155\u2013163 (1993)","journal-title":"SIAM J. Optim."},{"key":"795_CR8","unstructured":"Needell, D., Srebro, N., Ward, R.: Stochastic gradient descent and the randomized Kaczmarz algorithm, arXiv:1310.5715v1 [math.NA] (2013)"},{"key":"795_CR9","first-page":"372","volume":"27","author":"Y Nesterov","year":"1983","unstructured":"Nesterov, Y.: A method of solving a convex programming problem with convergence rate O(1\/ $$k^2$$ k 2 ). Sov. Math. Doklady 27, 372\u2013376 (1983)","journal-title":"Sov. Math. Doklady"},{"key":"795_CR10","doi-asserted-by":"crossref","DOI":"10.1007\/978-1-4419-8853-9","volume-title":"Introductory Lectures on Convex Optimization: A Basic Course","author":"Y Nesterov","year":"2004","unstructured":"Nesterov, Y.: Introductory Lectures on Convex Optimization: A Basic Course. Kluwer Academic Publishers, London (2004)"},{"key":"795_CR11","unstructured":"Nesterov, Y.: Gradient methods for minimizing composite objective function, CORE discussion paper (2007)"},{"issue":"3","key":"795_CR12","first-page":"123","volume":"1","author":"N Parikh","year":"2014","unstructured":"Parikh, N., Boyd, S.: Proximal algorithm. Found. Trends Optim. 1(3), 123\u2013231 (2014)","journal-title":"Found. Trends Optim."},{"key":"795_CR13","unstructured":"Man-Cho So, A.: Non-asymptotic convergence analysis of inexact gradient methods for machine learning without strong convexity, arXiv:1309.0113v1 [math.OC] (2013)"},{"key":"795_CR14","doi-asserted-by":"crossref","first-page":"425","DOI":"10.1007\/BF00941397","volume":"71","author":"P Tseng","year":"1991","unstructured":"Tseng, P.: Descent methods for convex essentially smooth minimization. J. Optim. Theory Appl. 71, 425\u2013463 (1991)","journal-title":"J. Optim. Theory Appl."},{"key":"795_CR15","doi-asserted-by":"crossref","DOI":"10.1007\/978-1-4757-2440-0","volume-title":"The Nature of Statistical Learning Theory","author":"V Vapnik","year":"1995","unstructured":"Vapnik, V.: The Nature of Statistical Learning Theory. Springer, New York (1995)"},{"key":"795_CR16","volume-title":"Iteration complexity of feasible descent methods for convex optimization, TR","author":"P Wang","year":"2013","unstructured":"Wang, P., Lin, C.: Iteration complexity of feasible descent methods for convex optimization, TR. National Taiwan Univerisity, Taipei (2013)"},{"issue":"1","key":"795_CR17","doi-asserted-by":"crossref","first-page":"143","DOI":"10.1137\/070703983","volume":"1","author":"W Yin","year":"2008","unstructured":"Yin, W., Osher, S., Goldfarb, D., Darbon, J.: Bregman iterative algorithms for $$\\ell _1$$ \u2113 1 -minimization with applications to compressed sensing. SIAM J. Imaging Sci. 1(1), 143\u2013168 (2008)","journal-title":"SIAM J. Imaging Sci."},{"key":"795_CR18","unstructured":"Zhang, H., Yin, W.: Gradient methods for convex minimization: better rates under weaker conditions, arXiv:1303.4645 , CAM Report 13\u201317, UCLA (2013)"},{"key":"795_CR19","doi-asserted-by":"crossref","unstructured":"Zhang, H., Cheng, L., Yin, W.: A dual algorithm for a class of augmented convex models, arXiv:1308.6337v1 , CAM Report 13\u201349, UCLA, 2013. Accepted by Communications in Mathematical Sciences (2014)","DOI":"10.4310\/CMS.2015.v13.n1.a5"}],"container-title":["Optimization Letters"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11590-014-0795-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s11590-014-0795-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11590-014-0795-x","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,8,15]],"date-time":"2019-08-15T00:31:09Z","timestamp":1565829069000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s11590-014-0795-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2014,9,16]]},"references-count":19,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2015,6]]}},"alternative-id":["795"],"URL":"https:\/\/doi.org\/10.1007\/s11590-014-0795-x","relation":{},"ISSN":["1862-4472","1862-4480"],"issn-type":[{"value":"1862-4472","type":"print"},{"value":"1862-4480","type":"electronic"}],"subject":[],"published":{"date-parts":[[2014,9,16]]}}}