{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T23:56:45Z","timestamp":1775001405954,"version":"3.50.1"},"publisher-location":"Cham","reference-count":22,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032157904","type":"print"},{"value":"9783032157911","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-15791-1_7","type":"book-chapter","created":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T22:58:24Z","timestamp":1774997904000},"page":"88-103","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["On the\u00a0Convergence of\u00a0First-Order Methods for\u00a0Quasar Convex Functions"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-1080-361X","authenticated-orcid":false,"given":"O. Santiago","family":"Valdivia Viscarra","sequence":"first","affiliation":[]},{"given":"Fedor","family":"Stonyakin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,4,1]]},"reference":[{"key":"7_CR1","doi-asserted-by":"crossref","unstructured":"Bayandina, A., Dvurechensky, P., Gasnikov, A., Stonyakin, F., Titov, A.: Mirror Descent and Convex Optimization Problems With Non-Smooth Inequality Constraints (2018)","DOI":"10.1007\/978-3-319-97478-1_8"},{"issue":"3\u20134","key":"7_CR2","doi-asserted-by":"publisher","first-page":"231","DOI":"10.1561\/2200000050","volume":"8","author":"S Bubeck","year":"2015","unstructured":"Bubeck, S.: Convex optimization: algorithms and complexity. Found. Trends Mach. Learn. 8(3\u20134), 231\u2013357 (2015)","journal-title":"Found. Trends Mach. Learn."},{"key":"7_CR3","first-page":"2121","volume":"12","author":"J Duchi","year":"2011","unstructured":"Duchi, J., Hazan, E., Singer, Y.: Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res. 12, 2121\u20132159 (2011)","journal-title":"J. Mach. Learn. Res."},{"key":"7_CR4","unstructured":"Fu, Q., Xu, D., Wilson, A.: Accelerated Stochastic Optimization Methods under Quasar-convexity. Proceedings of the 40-th International Conference on Machine Learning (2023)"},{"key":"7_CR5","unstructured":"Gasnikov, A.: Sovremennye chislennye metody optimizatsii metod universalnogo gradientnogo spuska (2018)"},{"key":"7_CR6","unstructured":"Gordon, G., Tibshirani, R.: Lecture Notes School of Computer Science. Carnegie Mellon University, (2012)"},{"key":"7_CR7","unstructured":"Hazan, E.: Introduction to Online Convex Optimization, The MIT Press (2022)"},{"key":"7_CR8","unstructured":"Hermant, J., Aujol, J., Dossal, C., Rondepierre, A.: Study of the behaviour of Nesterov Accelerated Gradient in a non convex setting: the strongly quasar convex case (2025)"},{"key":"7_CR9","unstructured":"Hinder, O., Sidford, A., Sohoni, N.: Near-Optimal Methods for Minimizing Star-Convex Functions and Beyond (2023)"},{"key":"7_CR10","unstructured":"Hardt, M., Ma, T., Recht, B.: Gradient Descent Learns Linear Dynamical Systems. J. Mach. Learn. Res. 19 (2018)"},{"key":"7_CR11","doi-asserted-by":"crossref","unstructured":"Huang, Y., Lin, Q.: Oracle complexity of single-loop switching subgradient methods for non-smooth weakly convex functional constrained optimization. In: 37th Conference on Neural Information Processing Systems (2023)","DOI":"10.52202\/075280-2679"},{"key":"7_CR12","unstructured":"Kleinberg, R., Li, Yuan, Y.: An alternative view: When does SGD escape local minima?. In International Conference on Machine Learning (ICML), pages 2698\u20132707, (2018)"},{"key":"7_CR13","volume-title":"On the Convergence of AdaGrad (norm) on Rd\u0302: Beyond Convexity","author":"Z Liu","year":"2023","unstructured":"Liu, Z., Nguyen, T., Ene, A., Nguyen, H.: On the Convergence of AdaGrad (norm) on Rd\u0302: Beyond Convexity. Published as a conference paper at ICLR, Non-asymptotic Rate and Acceleration (2023)"},{"key":"7_CR14","volume-title":"Problem Complexity and Method Efficiency in Optimization","author":"A Nemirovsky","year":"1983","unstructured":"Nemirovsky, A., Yudin, D.: Problem Complexity and Method Efficiency in Optimization. J. Wiley & Sons, New York (1983)"},{"key":"7_CR15","unstructured":"Polyak, B.: A General method for solving Extremum problems. Doklady Akademii nauk (1966)"},{"key":"7_CR16","unstructured":"Pun, Y.-M., Shames, I.: Online Non-Stationary Stochastic Quasar-Convex Optimization (2024)"},{"key":"7_CR17","unstructured":"Stich, S.: Universality of AdaGrad Stepsizes for Stochastic Optimization: Inexact Oracle, Acceleration and Variance Reduction (2024)"},{"key":"7_CR18","doi-asserted-by":"crossref","unstructured":"Titov, A., Stonyakin, F., Alkousa, M., Ablaev, S., Gasnikov, A.: Analogues of Switching Subgradient Schemes for Relatively Lipschitz-Continuous Convex Programming Problems (2021)","DOI":"10.1007\/978-3-030-58657-7_13"},{"key":"7_CR19","unstructured":"Titov, A., Stonyakin, F., Gasnikov, A., Alkousa, M.: Mirror Descent and Constrained Online Optimization Problems (2021)"},{"key":"7_CR20","unstructured":"Vorontsova, E., Hildebrand, R., Gasnikov, A., Stonyakin, F.: Convex Optimization (2021)"},{"key":"7_CR21","unstructured":"Wan, J.-K., Wibisono, A.: Continuized Acceleration for Quasar Convex Functions in Non-convex Optimization. Published as a conference paper at ICLR (2023)"},{"key":"7_CR22","unstructured":"Zhou, Y., Yang, J., Zhang, H., Liang, Y., Tarokh, V.: SGD converges to global minimum in deep learning via star-convex path. In: International Conference on Learning Representations (ICLR) (2019)"}],"container-title":["Lecture Notes in Computer Science","Optimization and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-15791-1_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T22:58:26Z","timestamp":1774997906000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-15791-1_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9783032157904","9783032157911"],"references-count":22,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-15791-1_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"1 April 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"OPTIMA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Optimization and Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Petrovac","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Montenegro","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 September 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 September 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"optima2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/conf-optima.ru","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}