{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T00:42:45Z","timestamp":1772757765052,"version":"3.50.1"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"1-3","license":[{"start":{"date-parts":[[2022,4,27]],"date-time":"2022-04-27T00:00:00Z","timestamp":1651017600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,4,27]],"date-time":"2022-04-27T00:00:00Z","timestamp":1651017600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"name":"Ivannikov Institute for System Programming of the Russian Academy of Sciences","award":["70-2021-00142"],"award-info":[{"award-number":["70-2021-00142"]}]},{"name":"Grant for research centers in the field of artificial intelligence, provided by the Analytical Center for the Government of the Russian Federation in accordance with the subsidy agreement","award":["000000D730321P5Q000"],"award-info":[{"award-number":["000000D730321P5Q000"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Optim Theory Appl"],"published-print":{"date-parts":[[2022,6]]},"DOI":"10.1007\/s10957-022-02038-7","type":"journal-article","created":{"date-parts":[[2022,4,27]],"date-time":"2022-04-27T12:04:43Z","timestamp":1651061083000},"page":"462-490","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Oracle Complexity Separation in Convex Optimization"],"prefix":"10.1007","volume":"193","author":[{"given":"Anastasiya","family":"Ivanova","sequence":"first","affiliation":[]},{"given":"Pavel","family":"Dvurechensky","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2173-6503","authenticated-orcid":false,"given":"Evgeniya","family":"Vorontsova","sequence":"additional","affiliation":[]},{"given":"Dmitry","family":"Pasechnyuk","sequence":"additional","affiliation":[]},{"given":"Alexander","family":"Gasnikov","sequence":"additional","affiliation":[]},{"given":"Darina","family":"Dvinskikh","sequence":"additional","affiliation":[]},{"given":"Alexander","family":"Tyurin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,27]]},"reference":[{"key":"2038_CR1","unstructured":"Agarwal, A., Bottou, L.: A lower bound for the optimization of finite sums. In: Bach, F., Blei, D. (eds.) Proceedings of the 32nd International Conference on Machine Learning, Proceedings of Machine Learning Research, vol.\u00a037, pp. 78\u201386. JMLR, Inc. and Microtome Publishing, Lille (2015). https:\/\/proceedings.mlr.press\/v37\/agarwal15.html"},{"key":"2038_CR2","doi-asserted-by":"crossref","unstructured":"Alkousa, M., Gasnikov, A., Dvurechensky, P., Sadiev, A., Razouk, L.: An Approach for Non-convex Uniformly Concave Structured Saddle Point Problem. arXiv:2202.06376 (2022)","DOI":"10.20537\/2076-7633-2022-14-2-225-237"},{"issue":"221","key":"2038_CR3","first-page":"1","volume":"18","author":"Z Allen-Zhu","year":"2018","unstructured":"Allen-Zhu, Z.: Katyusha: the first direct acceleration of stochastic gradient methods. J. Mach. Learn. Res. 18(221), 1\u201351 (2018)","journal-title":"J. Mach. Learn. Res."},{"key":"2038_CR4","unstructured":"Allen-Zhu, Z., Qu, Z., Richt\u00e1rik, P., Yuan, Y.: Even faster accelerated coordinate descent using non-uniform sampling. In: Balcan, M.F., Weinberger, K.Q. (eds.) Proceedings of The 33rd International Conference on Machine Learning, Proceedings of Machine Learning Research, vol.\u00a048, pp. 1110\u20131119. JMLR, Inc. and Microtome Publishing, New York. http:\/\/proceedings.mlr.press\/v48\/allen-zhuc16.html (2016)"},{"issue":"2","key":"2038_CR5","doi-asserted-by":"publisher","first-page":"4038","DOI":"10.1016\/j.ifacol.2020.12.2272","volume":"53","author":"A Beznosikov","year":"2020","unstructured":"Beznosikov, A., Gorbunov, E., Gasnikov, A.: Derivative-free method for composite optimization with applications to decentralized distributed optimization. IFAC-PapersOnLine 53(2), 4038\u20134043 (2020)","journal-title":"IFAC-PapersOnLine"},{"key":"2038_CR6","unstructured":"Bogolubsky, L., Dvurechenskii, P., Gasnikov, A., Gusev, G., Nesterov, Y., Raigorodskii, A.M., Tikhonov, A., Zhukovskii, M.: Learning supervised pagerank with gradient-based and gradient-free optimization methods. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds) Advances in Neural Information Processing Systems, vol.\u00a029, pp. 4914-4922. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2016\/file\/1f34004ebcb05f9acda6016d5cc52d5e-Paper.pdf (2016)"},{"key":"2038_CR7","doi-asserted-by":"publisher","unstructured":"Chen, P.Y., Zhang, H., Sharma, Y., Yi, J., Hsieh, C.J.: ZOO: Zeroth Order Optimization Based Black-Box Attacks to Deep Neural Networks without Training Substitute Models, pp. 15\u201326. Association for Computing Machinery, New York. https:\/\/doi.org\/10.1145\/3128572.3140448 (2017)","DOI":"10.1145\/3128572.3140448"},{"issue":"3","key":"2038_CR8","doi-asserted-by":"publisher","first-page":"385","DOI":"10.1515\/jiip-2020-0068","volume":"29","author":"D Dvinskikh","year":"2021","unstructured":"Dvinskikh, D., Gasnikov, A.: Decentralized and parallel primal and dual accelerated methods for stochastic convex programming problems. J. Inverse Ill-Posed Probl. 29(3), 385\u2013405 (2021). https:\/\/doi.org\/10.1515\/jiip-2020-0068","journal-title":"J. Inverse Ill-Posed Probl."},{"key":"2038_CR9","unstructured":"Dvurechensky, P., Gasnikov, A., Tiurin, A., Zholobov, V.: Unifying framework for accelerated randomized methods in convex optimization. arXiv:1707.08486 (2017)"},{"issue":"2","key":"2038_CR10","doi-asserted-by":"publisher","first-page":"601","DOI":"10.1016\/j.ejor.2020.08.027","volume":"290","author":"P Dvurechensky","year":"2021","unstructured":"Dvurechensky, P., Gorbunov, E., Gasnikov, A.: An accelerated directional derivative method for smooth stochastic convex optimization. Eur. J. Oper. Res. 290(2), 601\u2013621 (2021). https:\/\/doi.org\/10.1016\/j.ejor.2020.08.027","journal-title":"Eur. J. Oper. Res."},{"key":"2038_CR11","doi-asserted-by":"publisher","unstructured":"Dvurechensky, P., Shtern, S., Staudigl, M.: First-order methods for convex optimization. EURO J. Comput. Optim. 9, 100015 (2021). https:\/\/doi.org\/10.1016\/j.ejco.2021.100015, https:\/\/www.sciencedirect.com\/science\/article\/pii\/S2192440621001428, arXiv:2101.00935","DOI":"10.1016\/j.ejco.2021.100015"},{"key":"2038_CR12","doi-asserted-by":"publisher","unstructured":"Dvurechensky, P.E., Gasnikov, A.V., Nurminski, E.A., Stonyakin, F.S.: Advances in low-memory subgradient optimization, pp. 19\u201359. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-34910-3_2, arXiv:1902.01572","DOI":"10.1007\/978-3-030-34910-3_2"},{"issue":"4","key":"2038_CR13","doi-asserted-by":"publisher","first-page":"1997","DOI":"10.1137\/130949993","volume":"25","author":"O Fercoq","year":"2015","unstructured":"Fercoq, O., Richt\u00e1rik, P.: Accelerated, parallel, and proximal coordinate descent. SIAM J. Optim. 25(4), 1997\u20132023 (2015). https:\/\/doi.org\/10.1137\/130949993","journal-title":"SIAM J. Optim."},{"issue":"2","key":"2038_CR14","first-page":"67","volume":"8","author":"A Gasnikov","year":"2016","unstructured":"Gasnikov, A., Dvurechensky, P., Usmanova, I.: About accelerated randomized methods. Proc. Moscow Inst. Phys. Technol. 8(2), 67\u2013100 (2016)","journal-title":"Proc. Moscow Inst. Phys. Technol."},{"key":"2038_CR15","unstructured":"Gasnikov, A., Novitskii, A., Novitskii, V., Abdukhakimov, F., Kamzolov, D., Beznosikov, A., Takac, M., Dvurechensky, P., Gu, B.: The power of first-order smooth optimization for black-box non-smooth problems. arXiv:2201.12289 (2022)"},{"issue":"1","key":"2038_CR16","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1134\/s096554252101005x","volume":"61","author":"AV Gasnikov","year":"2021","unstructured":"Gasnikov, A.V., Dvinskikh, D.M., Dvurechensky, P.E., Kamzolov, D.I., Matyukhin, V.V., Pasechnyuk, D.A., Tupitsa, N.K., Chernov, A.V.: Accelerated meta-algorithm for convex optimization problems. Comput. Math. Math. Phys. 61(1), 17\u201328 (2021). https:\/\/doi.org\/10.1134\/s096554252101005x","journal-title":"Comput. Math. Math. Phys."},{"key":"2038_CR17","doi-asserted-by":"crossref","unstructured":"Gladin, E., Sadiev, A., Gasnikov, A., Dvurechensky, P., Beznosikov, A., Alkousa, M.: Solving smooth min-min and min-max problems by mixed oracle algorithms. In: Strekalovsky, A., Kochetov, Y., Gruzdeva, T., Orlov, A. (eds) Mathematical Optimization Theory and Operations Research: Recent Trends, pp. 19\u201340. Springer, Cham (2021). https:\/\/link.springer.com\/chapter\/10.1007\/978-3-030-86433-0_2","DOI":"10.1007\/978-3-030-86433-0_2"},{"key":"2038_CR18","doi-asserted-by":"crossref","unstructured":"Gorbunov, E., Dvurechensky, P., Gasnikov, A.: An accelerated method for derivative-free smooth stochastic convex optimization. SIAM J. Optim. (2022). (accepted). arXiv:1802.09022","DOI":"10.1016\/j.ejor.2020.08.027"},{"key":"2038_CR19","doi-asserted-by":"publisher","unstructured":"Ivanova, A., Pasechnyuk, D., Grishchenko, D., Shulgin, E., Gasnikov, A., Matyukhin, V.: Adaptive catalyst for smooth convex optimization. In: Optimization and Applications, pp. 20\u201337. Springer (2021). https:\/\/doi.org\/10.1007\/978-3-030-91059-4_2","DOI":"10.1007\/978-3-030-91059-4_2"},{"key":"2038_CR20","unstructured":"Ivanova, A., Vorontsova, E., Pasechnyuk, D., Gasnikov, A., Dvurechensky, P., Dvinskikh, D., Tyurin, A.: Oracle complexity separation in convex optimization. arXiv:2002.02706 (2020)"},{"key":"2038_CR21","first-page":"166","volume-title":"Optim. Appl.","author":"D Kamzolov","year":"2020","unstructured":"Kamzolov, D., Gasnikov, A., Dvurechensky, P.: Optimal combination of tensor optimization methods. In: Olenev, N., Evtushenko, Y., Khachay, M., Malkova, V. (eds.) Optim. Appl., pp. 166\u2013183. Springer, Cham (2020)"},{"issue":"1\u20132","key":"2038_CR22","doi-asserted-by":"publisher","first-page":"201","DOI":"10.1007\/s10107-015-0955-5","volume":"159","author":"G Lan","year":"2015","unstructured":"Lan, G.: Gradient sliding for composite optimization. Math. Program. 159(1\u20132), 201\u2013235 (2015). https:\/\/doi.org\/10.1007\/s10107-015-0955-5","journal-title":"Math. Program."},{"key":"2038_CR23","unstructured":"Lan, G., Li, Z., Zhou, Y.: A unified variance-reduced accelerated gradient method for convex optimization. In: Advances in Neural Information Processing Systems, pp. 10462\u201310472. Curran Associates Inc. (2019)"},{"key":"2038_CR24","doi-asserted-by":"crossref","unstructured":"Lan, G., Ouyang, Y.: Accelerated gradient sliding for structured convex optimization. arXiv:1609.04905 (2016)","DOI":"10.1137\/140992382"},{"key":"2038_CR25","unstructured":"Lan, G., Ouyang, Y.: Mirror-prox sliding methods for solving a class of monotone variational inequalities. arXiv:2111.00996 (2021)"},{"issue":"2","key":"2038_CR26","doi-asserted-by":"publisher","first-page":"1379","DOI":"10.1137\/140992382","volume":"26","author":"G Lan","year":"2016","unstructured":"Lan, G., Zhou, Y.: Conditional gradient sliding for convex optimization. SIAM J. Optim. 26(2), 1379\u20131409 (2016). https:\/\/doi.org\/10.1137\/140992382","journal-title":"SIAM J. Optim."},{"issue":"1\u20132","key":"2038_CR27","doi-asserted-by":"publisher","first-page":"167","DOI":"10.1007\/s10107-017-1173-0","volume":"171","author":"G Lan","year":"2017","unstructured":"Lan, G., Zhou, Y.: An optimal randomized incremental gradient method. Math. Program. 171(1\u20132), 167\u2013215 (2017). https:\/\/doi.org\/10.1007\/s10107-017-1173-0","journal-title":"Math. Program."},{"key":"2038_CR28","unstructured":"Lin, H., Mairal, J., Harchaoui, Z.: A universal catalyst for first-order optimization. In: Cortes, C., Lawrence, N., Lee, D., Sugiyama, M., Garnett, R. (eds) Advances in Neural Information Processing Systems, vol.\u00a028, pp. 3384\u20133392. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2015\/file\/c164bbc9d6c72a52c599bbb43d8db8e1-Paper.pdf (2015)"},{"key":"2038_CR29","unstructured":"Lin, H., Mairal, J., Harchaoui, Z.: Catalyst acceleration for first-order convex optimization: from theory to practice. J. Mach. Learn. Res. 18(212), 1\u201354 (2018). http:\/\/jmlr.org\/papers\/v18\/17-748.html"},{"key":"2038_CR30","unstructured":"Lin, Q., Lu, Z., Xiao, L.: An accelerated proximal coordinate gradient method. In: Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., Weinberger, K.Q. (eds) Advances in Neural Information Processing Systems, vol.\u00a027, pp. 3059\u20133067. Curran Associates, Inc. (2014). https:\/\/proceedings.neurips.cc\/paper\/2014\/file\/8f19793b2671094e63a15ab883d50137-Paper.pdf"},{"issue":"2","key":"2038_CR31","doi-asserted-by":"publisher","first-page":"1092","DOI":"10.1137\/110833786","volume":"23","author":"RDC Monteiro","year":"2013","unstructured":"Monteiro, R.D.C., Svaiter, B.F.: An accelerated hybrid proximal extragradient method for convex optimization and its implications to second-order methods. SIAM J. Optim. 23(2), 1092\u20131125 (2013). https:\/\/doi.org\/10.1137\/110833786","journal-title":"SIAM J. Optim."},{"key":"2038_CR32","volume-title":"Problem Complexity and Method Efficiency in Optimization","author":"AS Nemirovsky","year":"1983","unstructured":"Nemirovsky, A.S., Yudin, D.B.: Problem Complexity and Method Efficiency in Optimization. Wiley-Blackwell, Chichester, New York (1983)"},{"issue":"1","key":"2038_CR33","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1007\/s10107-004-0552-5","volume":"103","author":"Y Nesterov","year":"2004","unstructured":"Nesterov, Y.: Smooth minimization of non-smooth functions. Math. Program. 103(1), 127\u2013152 (2004). https:\/\/doi.org\/10.1007\/s10107-004-0552-5","journal-title":"Math. Program."},{"issue":"2","key":"2038_CR34","doi-asserted-by":"publisher","first-page":"341","DOI":"10.1137\/100802001","volume":"22","author":"Y Nesterov","year":"2012","unstructured":"Nesterov, Y.: Efficiency of coordinate descent methods on huge-scale optimization problems. SIAM J. Optim. 22(2), 341\u2013362 (2012). https:\/\/doi.org\/10.1137\/100802001","journal-title":"SIAM J. Optim."},{"issue":"1","key":"2038_CR35","doi-asserted-by":"publisher","first-page":"125","DOI":"10.1007\/s10107-012-0629-5","volume":"140","author":"Y Nesterov","year":"2012","unstructured":"Nesterov, Y.: Gradient methods for minimizing composite functions. Math. Program. 140(1), 125\u2013161 (2012). https:\/\/doi.org\/10.1007\/s10107-012-0629-5","journal-title":"Math. Program."},{"key":"2038_CR36","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-91578-4","volume-title":"Lectures on Convex Optimization","author":"Y Nesterov","year":"2018","unstructured":"Nesterov, Y.: Lectures on Convex Optimization, 2nd edn. Springer, Berlin (2018)","edition":"2"},{"issue":"2","key":"2038_CR37","doi-asserted-by":"publisher","first-page":"527","DOI":"10.1007\/s10208-015-9296-2","volume":"17","author":"Y Nesterov","year":"2015","unstructured":"Nesterov, Y., Spokoiny, V.: Random gradient-free minimization of convex functions. Found. Comput. Math. 17(2), 527\u2013566 (2015). https:\/\/doi.org\/10.1007\/s10208-015-9296-2","journal-title":"Found. Comput. Math."},{"issue":"1","key":"2038_CR38","doi-asserted-by":"publisher","first-page":"110","DOI":"10.1137\/16M1060182","volume":"27","author":"Y Nesterov","year":"2017","unstructured":"Nesterov, Y., Stich, S.U.: Efficiency of the accelerated coordinate descent method on structured optimization problems. SIAM J. Optim. 27(1), 110\u2013123 (2017). https:\/\/doi.org\/10.1137\/16M1060182","journal-title":"SIAM J. Optim."},{"key":"2038_CR39","unstructured":"Rogozin, A., Beznosikov, A., Dvinskikh, D., Kovalev, D., Dvurechensky, P., Gasnikov, A.: Decentralized distributed optimization for saddle point problems. arXiv:2102.07758 (2021)"},{"key":"2038_CR40","doi-asserted-by":"publisher","unstructured":"Rogozin, A., Bochko, M., Dvurechensky, P., Gasnikov, A., Lukoshkin, V.: An accelerated method for decentralized distributed stochastic optimization over time-varying graphs. In: 2021 60th IEEE Conference on Decision and Control (CDC), pp. 3367\u20133373. https:\/\/doi.org\/10.1109\/CDC45484.2021.9683110 (2021)","DOI":"10.1109\/CDC45484.2021.9683110"},{"key":"2038_CR41","doi-asserted-by":"crossref","unstructured":"Sadiev, A., Beznosikov, A., Dvurechensky, P., Gasnikov, A.: Zeroth-order algorithms for smooth saddle-point problems. In: Strekalovsky, A., Kochetov, Y., Gruzdeva, T., Orlov, A. (eds) Mathematical Optimization Theory and Operations Research: Recent Trends, pp. 71\u201385. Springer, Cham (2021). https:\/\/link.springer.com\/chapter\/10.1007\/978-3-030-86433-0_5, ArXiv:2009.09908","DOI":"10.1007\/978-3-030-86433-0_5"},{"key":"2038_CR42","doi-asserted-by":"crossref","DOI":"10.7551\/mitpress\/4175.001.0001","volume-title":"Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond","author":"B Sch\u00f6lkopf","year":"2001","unstructured":"Sch\u00f6lkopf, B., Smola, A.J.: Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond. MIT Press, Cambridge (2001)"},{"key":"2038_CR43","doi-asserted-by":"publisher","DOI":"10.1007\/s11590-021-01742-z","author":"I Shibaev","year":"2021","unstructured":"Shibaev, I., Dvurechensky, P., Gasnikov, A.: Zeroth-order methods for noisy H\u00f6lder-gradient functions. Optim. Lett. (2021). https:\/\/doi.org\/10.1007\/s11590-021-01742-z","journal-title":"Optim. Lett."},{"key":"2038_CR44","unstructured":"Spokoiny, V., Panov, M.: Accuracy of gaussian approximation in nonparametric Bernstein\u2013von Mises theorem. arXiv:1910.06028 (2019)"},{"key":"2038_CR45","unstructured":"Stepanov, I., Voronov, A., Beznosikov, A., Gasnikov, A.: One-point gradient-free methods for composite optimization with applications to distributed optimization. arXiv:2107.05951 (2021)"},{"key":"2038_CR46","doi-asserted-by":"publisher","DOI":"10.1080\/10556788.2021.1924714","author":"F Stonyakin","year":"2021","unstructured":"Stonyakin, F., Tyurin, A., Gasnikov, A., Dvurechensky, P., Agafonov, A., Dvinskikh, D., Alkousa, M., Pasechnyuk, D., Artamonov, S., Piskunova, V.: Inexact model: a framework for optimization and variational inequalities. Optim. Methods Softw. (2021). https:\/\/doi.org\/10.1080\/10556788.2021.1924714","journal-title":"Optim. Methods Softw."},{"key":"2038_CR47","unstructured":"Tominin, V., Tominin, Y., Borodich, E., Kovalev, D., Gasnikov, A., Dvurechensky, P.: On accelerated methods for saddle-point problems with composite structure. arXiv:2103.09344 (2021)"},{"key":"2038_CR48","doi-asserted-by":"crossref","unstructured":"Tu, C.C., Ting, P., Chen, P.Y., Liu, S., Zhang, H., Yi, J., Hsieh, C.J., Cheng, S.M.: Autozoom: autoencoder-based zeroth order optimization method for attacking black-box neural networks. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a033, pp. 742\u2013749 (2019)","DOI":"10.1609\/aaai.v33i01.3301742"},{"key":"2038_CR49","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4757-2440-0","volume-title":"The Nature of Statistical Learning Theory","author":"VN Vapnik","year":"1995","unstructured":"Vapnik, V.N.: The Nature of Statistical Learning Theory. Springer, New York (1995)"},{"issue":"8","key":"2038_CR50","doi-asserted-by":"publisher","first-page":"1487","DOI":"10.1134\/s0005117919080095","volume":"80","author":"EA Vorontsova","year":"2019","unstructured":"Vorontsova, E.A., Gasnikov, A.V., Gorbunov, E.A., Dvurechenskii, P.E.: Accelerated gradient-free optimization methods with a non-Euclidean proximal operator. Autom. Remote. Control. 80(8), 1487\u20131501 (2019). https:\/\/doi.org\/10.1134\/s0005117919080095","journal-title":"Autom. Remote. Control."},{"key":"2038_CR51","unstructured":"Zhang, X., Saha, A., Vishwanathan, S.: Regularized risk minimization by Nesterov\u2019s accelerated gradient methods: algorithmic extensions and empirical studies. arXiv:1011.0472 (2010)"},{"key":"2038_CR52","unstructured":"Zhang, Y., Xiao, L.: Stochastic primal-dual coordinate method for regularized empirical risk minimization. In: Bach, F., Blei, D. (eds) Proceedings of the 32nd International Conference on Machine Learning, Proceedings of Machine Learning Research, vol.\u00a037, pp. 353\u2013361. PMLR, Lille. http:\/\/proceedings.mlr.press\/v37\/zhanga15.html (2015)"}],"container-title":["Journal of Optimization Theory and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10957-022-02038-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10957-022-02038-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10957-022-02038-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,23]],"date-time":"2024-09-23T02:22:38Z","timestamp":1727058158000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10957-022-02038-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,27]]},"references-count":52,"journal-issue":{"issue":"1-3","published-print":{"date-parts":[[2022,6]]}},"alternative-id":["2038"],"URL":"https:\/\/doi.org\/10.1007\/s10957-022-02038-7","relation":{},"ISSN":["0022-3239","1573-2878"],"issn-type":[{"value":"0022-3239","type":"print"},{"value":"1573-2878","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,27]]},"assertion":[{"value":"31 March 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 April 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 April 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}