{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,8]],"date-time":"2026-02-08T03:40:05Z","timestamp":1770522005061,"version":"3.49.0"},"reference-count":64,"publisher":"Springer Science and Business Media LLC","issue":"1-2","license":[{"start":{"date-parts":[[2023,8,16]],"date-time":"2023-08-16T00:00:00Z","timestamp":1692144000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,8,16]],"date-time":"2023-08-16T00:00:00Z","timestamp":1692144000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Math. Program."],"published-print":{"date-parts":[[2024,9]]},"DOI":"10.1007\/s10107-023-02008-5","type":"journal-article","created":{"date-parts":[[2023,8,16]],"date-time":"2023-08-16T16:02:22Z","timestamp":1692201742000},"page":"227-268","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["A new complexity metric for nonconvex rank-one generalized matrix completion"],"prefix":"10.1007","volume":"207","author":[{"given":"Haixiang","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Baturalp","family":"Yalcin","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4294-1338","authenticated-orcid":false,"given":"Javad","family":"Lavaei","sequence":"additional","affiliation":[]},{"given":"Somayeh","family":"Sojoudi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,16]]},"reference":[{"issue":"4","key":"2008_CR1","doi-asserted-by":"publisher","first-page":"2775","DOI":"10.1137\/140979861","volume":"26","author":"A Agarwal","year":"2016","unstructured":"Agarwal, A., Anandkumar, A., Jain, P., Netrapalli, P.: Learning sparsely used overcomplete dictionaries via alternating minimization. SIAM J. Optim. 26(4), 2775\u20132799 (2016)","journal-title":"SIAM J. Optim."},{"key":"2008_CR2","unstructured":"Ahn, K., Suarez, F.: Riemannian perspective on matrix factorization. arXiv preprint arXiv:2102.00937 (2021)"},{"issue":"9","key":"2008_CR3","first-page":"808","volume":"102","author":"M Aigner","year":"1995","unstructured":"Aigner, M.: Tur\u00e1n\u2019s graph theorem. Am. Math. Mon. 102(9), 808\u2013816 (1995)","journal-title":"Am. Math. Mon."},{"key":"2008_CR4","unstructured":"Ajayi, T., Mildebrath, D., Kyrillidis, A., Ubaru, S., Kollias, G., Bouchard, K.: Provably convergent acceleration in factored gradient descent with applications in matrix sensing. arXiv preprint arXiv:1806.00534 (2018)"},{"key":"2008_CR5","unstructured":"Allen-Zhu, Z., Li, Y.: Neon2: Finding local minima via first-order oracles. Adv. Neural Inform. Process. Syst. 31 (2018)"},{"key":"2008_CR6","unstructured":"Bi, Y., Lavaei, J.: On the absence of spurious local minima in nonlinear low-rank matrix recovery problems. In: International Conference on Artificial Intelligence and Statistics, pp. 379\u2013387. PMLR (2021)"},{"key":"2008_CR7","doi-asserted-by":"crossref","unstructured":"Bi, Y., Zhang, H., Lavaei, J.: Local and global linear convergence of general low-rank matrix recovery problems. In: Proceedings of 36th AAAI Conference on Artificial Intelligence (AAAI), pp. 1\u20139. Vancouver, Canada (2022)","DOI":"10.1609\/aaai.v36i9.21252"},{"issue":"2","key":"2008_CR8","doi-asserted-by":"publisher","first-page":"329","DOI":"10.1007\/s10107-002-0352-8","volume":"95","author":"S Burer","year":"2003","unstructured":"Burer, S., Monteiro, R.D.: A nonlinear programming algorithm for solving semidefinite programs via low-rank factorization. Math. Program. 95(2), 329\u2013357 (2003)","journal-title":"Math. Program."},{"issue":"3","key":"2008_CR9","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/1970392.1970395","volume":"58","author":"EJ Cand\u00e8s","year":"2011","unstructured":"Cand\u00e8s, E.J., Li, X., Ma, Y., Wright, J.: Robust principal component analysis? J. ACM (JACM) 58(3), 1\u201337 (2011)","journal-title":"J. ACM (JACM)"},{"issue":"4","key":"2008_CR10","doi-asserted-by":"publisher","first-page":"1985","DOI":"10.1109\/TIT.2015.2399924","volume":"61","author":"EJ Candes","year":"2015","unstructured":"Candes, E.J., Li, X., Soltanolkotabi, M.: Phase retrieval via wirtinger flow: theory and algorithms. IEEE Trans. Inform. Theory 61(4), 1985\u20132007 (2015)","journal-title":"IEEE Trans. Inform. Theory"},{"issue":"4","key":"2008_CR11","doi-asserted-by":"publisher","first-page":"2342","DOI":"10.1109\/TIT.2011.2111771","volume":"57","author":"EJ Candes","year":"2011","unstructured":"Candes, E.J., Plan, Y.: Tight oracle inequalities for low-rank matrix recovery from a minimal number of noisy random measurements. IEEE Trans. Inform. Theory 57(4), 2342\u20132359 (2011)","journal-title":"IEEE Trans. Inform. Theory"},{"issue":"6","key":"2008_CR12","doi-asserted-by":"publisher","first-page":"717","DOI":"10.1007\/s10208-009-9045-5","volume":"9","author":"EJ Cand\u00e8s","year":"2009","unstructured":"Cand\u00e8s, E.J., Recht, B.: Exact matrix completion via convex optimization. Found. Comput. Math. 9(6), 717\u2013772 (2009)","journal-title":"Found. Comput. Math."},{"issue":"5","key":"2008_CR13","doi-asserted-by":"publisher","first-page":"2053","DOI":"10.1109\/TIT.2010.2044061","volume":"56","author":"EJ Cand\u00e8s","year":"2010","unstructured":"Cand\u00e8s, E.J., Tao, T.: The power of convex relaxation: near-optimal matrix completion. IEEE Trans. Inform. Theory 56(5), 2053\u20132080 (2010)","journal-title":"IEEE Trans. Inform. Theory"},{"issue":"2","key":"2008_CR14","doi-asserted-by":"publisher","first-page":"245","DOI":"10.1007\/s10107-009-0286-5","volume":"127","author":"C Cartis","year":"2011","unstructured":"Cartis, C., Gould, N.I., Toint, P.L.: Adaptive cubic regularisation methods for unconstrained optimization. Part I: motivation, convergence and numerical results. Math. Program. 127(2), 245\u2013295 (2011)","journal-title":"Math. Program."},{"issue":"6","key":"2008_CR15","doi-asserted-by":"publisher","first-page":"1505","DOI":"10.1007\/s10208-020-09490-9","volume":"21","author":"V Charisopoulos","year":"2021","unstructured":"Charisopoulos, V., Chen, Y., Davis, D., D\u00edaz, M., Ding, L., Drusvyatskiy, D.: Low-rank matrix recovery with composite optimization: good conditioning and rapid convergence. Found. Comput. Math. 21(6), 1505\u20131593 (2021)","journal-title":"Found. Comput. Math."},{"issue":"142","key":"2008_CR16","first-page":"1","volume":"20","author":"J Chen","year":"2019","unstructured":"Chen, J., Li, X.: Model-free nonconvex matrix completion: local minima analysis and applications in memory-efficient kernel PCA. J. Mach. Learn. Res. 20(142), 1\u201339 (2019)","journal-title":"J. Mach. Learn. Res."},{"issue":"9","key":"2008_CR17","doi-asserted-by":"publisher","first-page":"5806","DOI":"10.1109\/TIT.2020.2992234","volume":"66","author":"J Chen","year":"2020","unstructured":"Chen, J., Liu, D., Li, X.: Nonconvex rectangular matrix completion via gradient descent without $$\\ell _{2,\\infty }$$ regularization. IEEE Trans. Inform. Theory 66(9), 5806\u20135841 (2020)","journal-title":"IEEE Trans. Inform. Theory"},{"issue":"4","key":"2008_CR18","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1109\/MSP.2018.2821706","volume":"35","author":"Y Chen","year":"2018","unstructured":"Chen, Y., Chi, Y.: Harnessing structures in big data via guaranteed low-rank matrix estimation: recent theory and fast algorithms via convex and nonconvex optimization. IEEE Signal Process. Mag. 35(4), 14\u201331 (2018)","journal-title":"IEEE Signal Process. Mag."},{"issue":"1","key":"2008_CR19","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1007\/s10107-019-01363-6","volume":"176","author":"Y Chen","year":"2019","unstructured":"Chen, Y., Chi, Y., Fan, J., Ma, C.: Gradient descent with random initialization: fast global convergence for nonconvex phase retrieval. Math. Program. 176(1), 5\u201337 (2019)","journal-title":"Math. Program."},{"issue":"4","key":"2008_CR20","doi-asserted-by":"publisher","first-page":"3098","DOI":"10.1137\/19M1290000","volume":"30","author":"Y Chen","year":"2020","unstructured":"Chen, Y., Chi, Y., Fan, J., Ma, C., Yan, Y.: Noisy matrix completion: understanding statistical guarantees for convex relaxation via nonconvex optimization. SIAM J. Optim. 30(4), 3098\u20133121 (2020)","journal-title":"SIAM J. Optim."},{"issue":"5","key":"2008_CR21","doi-asserted-by":"publisher","first-page":"2948","DOI":"10.1214\/21-AOS2066","volume":"49","author":"Y Chen","year":"2021","unstructured":"Chen, Y., Fan, J., Ma, C., Yan, Y.: Bridging convex and nonconvex optimization in robust PCA: noise, outliers and missing data. Ann. Stat. 49(5), 2948\u20132971 (2021)","journal-title":"Ann. Stat."},{"issue":"20","key":"2008_CR22","doi-asserted-by":"publisher","first-page":"5239","DOI":"10.1109\/TSP.2019.2937282","volume":"67","author":"Y Chi","year":"2019","unstructured":"Chi, Y., Lu, Y.M., Chen, Y.: Nonconvex optimization meets low-rank matrix factorization: an overview. IEEE Trans. Signal Process. 67(20), 5239\u20135269 (2019)","journal-title":"IEEE Trans. Signal Process."},{"key":"2008_CR23","unstructured":"Chou, H.H., Gieshoff, C., Maly, J., Rauhut, H.: Gradient descent for deep matrix factorization: dynamics and implicit bias towards low rank. arXiv preprint arXiv:2011.13772 (2020)"},{"key":"2008_CR24","unstructured":"Fattahi, S., Sojoudi, S.: Exact guarantees on the absence of spurious local minima for non-negative rank-1 robust principal component analysis. J. Mach. Learn. Res. (2020)"},{"key":"2008_CR25","unstructured":"Ge, R., Jin, C., Zheng, Y.: No spurious local minima in nonconvex low rank problems: a unified geometric analysis. In: International Conference on Machine Learning, pp. 1233\u20131242. PMLR (2017)"},{"key":"2008_CR26","unstructured":"Ge, R., Lee, J.D., Ma, T.: Matrix completion has no spurious local minimum. Adv. Neural Inform. Process. Syst. 2981\u20132989 (2016)"},{"key":"2008_CR27","doi-asserted-by":"crossref","unstructured":"Hardt, M.: Understanding alternating minimization for matrix completion. In: 2014 IEEE 55th Annual Symposium on Foundations of Computer Science, pp. 651\u2013660. IEEE (2014)","DOI":"10.1109\/FOCS.2014.75"},{"key":"2008_CR28","unstructured":"Hardt, M., Wootters, M.: Fast matrix completion without the condition number. In: Conference on Learning Theory, pp. 638\u2013678. PMLR (2014)"},{"key":"2008_CR29","unstructured":"Hou, T.Y., Li, Z., Zhang, Z.: Fast global convergence for low-rank matrix recovery via riemannian gradient descent with random initialization. arXiv preprint arXiv:2012.15467 (2020)"},{"key":"2008_CR30","unstructured":"Jain, P., Meka, R., Dhillon, I.: Guaranteed rank minimization via singular value projection. Adv. Neural Inform. Process. Syst. 23 (2010)"},{"key":"2008_CR31","doi-asserted-by":"crossref","unstructured":"Jain, P., Netrapalli, P., Sanghavi, S.: Low-rank matrix completion using alternating minimization. In: Proceedings of the Forty-fifth Annual ACM Symposium on Theory of Computing, pp. 665\u2013674. (2013)","DOI":"10.1145\/2488608.2488693"},{"key":"2008_CR32","unstructured":"Jin, C., Ge, R., Netrapalli, P., Kakade, S.M., Jordan, M.I.: How to escape saddle points efficiently. In: International Conference on Machine Learning, pp. 1724\u20131732. PMLR (2017)"},{"key":"2008_CR33","unstructured":"Jin, C., Netrapalli, P., Jordan, M.I.: Accelerated gradient descent escapes saddle points faster than gradient descent. In: Conference On Learning Theory, pp. 1042\u20131085. PMLR (2018)"},{"key":"2008_CR34","unstructured":"Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient descent only converges to minimizers. In: Conference on Learning Theory, pp. 1246\u20131257. PMLR (2016)"},{"key":"2008_CR35","unstructured":"Levin, E., Kileel, J., Boumal, N.: The effect of smooth parametrizations on nonconvex optimization landscapes. arXiv preprint arXiv:2207.03512 (2022)"},{"issue":"1","key":"2008_CR36","doi-asserted-by":"publisher","first-page":"660","DOI":"10.1137\/18M1224738","volume":"30","author":"X Li","year":"2020","unstructured":"Li, X., Zhu, Z., Man-Cho So, A., Vidal, R.: Nonconvex robust low-rank matrix recovery. SIAM J. Optim. 30(1), 660\u2013686 (2020)","journal-title":"SIAM J. Optim."},{"key":"2008_CR37","unstructured":"Li, Y., Ma, T., Zhang, H.: Algorithmic regularization in over-parameterized matrix sensing and neural networks with quadratic activations. In: Conference On Learning Theory, pp. 2\u201347. PMLR (2018)"},{"key":"2008_CR38","unstructured":"Luo, Y., Li, X., Zhang, A.R.: Nonconvex factorization and manifold formulations are almost equivalent in low-rank matrix optimization. arXiv preprint arXiv:2108.01772 (2021)"},{"key":"2008_CR39","unstructured":"Ma, J., Fattahi, S.: Sign-rip: a robust restricted isometry property for low-rank matrix recovery. arXiv preprint arXiv:2102.02969 (2021)"},{"key":"2008_CR40","doi-asserted-by":"crossref","unstructured":"Ma, Z., Bi, Y., Lavaei, J., Sojoudi, S.: Sharp restricted isometry property bounds for low-rank matrix recovery problems with corrupted measurements. arXiv preprint arXiv:2105.08232 (2021)","DOI":"10.1609\/aaai.v36i7.20734"},{"key":"2008_CR41","unstructured":"Netrapalli, P., Jain, P., Sanghavi, S.: Phase retrieval using alternating minimization. Adv. Neural Inform. Process. Syst. 26 (2013)"},{"key":"2008_CR42","unstructured":"Netrapalli, P., Un, N., Sanghavi, S., Anandkumar, A., Jain, P.: Non-convex robust PCA. Adv. Neural Inform. Process. Syst. 27 (2014)"},{"issue":"3","key":"2008_CR43","doi-asserted-by":"publisher","first-page":"471","DOI":"10.1137\/070697835","volume":"52","author":"B Recht","year":"2010","unstructured":"Recht, B., Fazel, M., Parrilo, P.A.: Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization. SIAM Rev. 52(3), 471\u2013501 (2010)","journal-title":"SIAM Rev."},{"issue":"1","key":"2008_CR44","doi-asserted-by":"publisher","first-page":"279","DOI":"10.1007\/BF01585941","volume":"70","author":"J Renegar","year":"1995","unstructured":"Renegar, J.: Linear programming, complexity theory and elementary functional analysis. Math. Program. 70(1), 279\u2013351 (1995)","journal-title":"Math. Program."},{"issue":"4","key":"2008_CR45","doi-asserted-by":"publisher","first-page":"879","DOI":"10.1137\/S105262349427532X","volume":"6","author":"J Renegar","year":"1996","unstructured":"Renegar, J.: Condition numbers, the barrier method, and the conjugate-gradient method. SIAM J. Optim. 6(4), 879\u2013912 (1996)","journal-title":"SIAM J. Optim."},{"key":"2008_CR46","unstructured":"St\u00f6ger, D., Soltanolkotabi, M.: Small random initialization is akin to spectral learning: optimization and generalization guarantees for overparameterized low-rank matrix reconstruction. Adv. Neural Inform. Process. Syst. 34 (2021)"},{"issue":"2","key":"2008_CR47","doi-asserted-by":"publisher","first-page":"853","DOI":"10.1109\/TIT.2016.2632162","volume":"63","author":"J Sun","year":"2016","unstructured":"Sun, J., Qu, Q., Wright, J.: Complete dictionary recovery over the sphere I: overview and the geometric picture. IEEE Trans. Inform. Theory 63(2), 853\u2013884 (2016)","journal-title":"IEEE Trans. Inform. Theory"},{"issue":"5","key":"2008_CR48","doi-asserted-by":"publisher","first-page":"1131","DOI":"10.1007\/s10208-017-9365-9","volume":"18","author":"J Sun","year":"2018","unstructured":"Sun, J., Qu, Q., Wright, J.: A geometric analysis of phase retrieval. Found. Comput. Math. 18(5), 1131\u20131198 (2018)","journal-title":"Found. Comput. Math."},{"issue":"11","key":"2008_CR49","doi-asserted-by":"publisher","first-page":"6535","DOI":"10.1109\/TIT.2016.2598574","volume":"62","author":"R Sun","year":"2016","unstructured":"Sun, R., Luo, Z.Q.: Guaranteed matrix completion via non-convex factorization. IEEE Trans. Inform. Theory 62(11), 6535\u20136579 (2016)","journal-title":"IEEE Trans. Inform. Theory"},{"issue":"150","key":"2008_CR50","first-page":"1","volume":"22","author":"T Tong","year":"2021","unstructured":"Tong, T., Ma, C., Chi, Y.: Accelerating ill-conditioned low-rank matrix estimation via scaled gradient descent. J. Mach. Learn. Res. 22(150), 1\u201363 (2021)","journal-title":"J. Mach. Learn. Res."},{"key":"2008_CR51","doi-asserted-by":"publisher","first-page":"2396","DOI":"10.1109\/TSP.2021.3071560","volume":"69","author":"T Tong","year":"2021","unstructured":"Tong, T., Ma, C., Chi, Y.: Low-rank matrix recovery with scaled subgradient methods: fast and robust convergence without the condition number. IEEE Trans. Signal Process. 69, 2396\u20132409 (2021)","journal-title":"IEEE Trans. Signal Process."},{"key":"2008_CR52","unstructured":"Tong, T., Ma, C., Prater-Bennette, A., Tripp, E., Chi, Y.: Scaling and scalability: provable nonconvex low-rank tensor estimation from incomplete measurements. arXiv preprint arXiv:2104.14526 (2021)"},{"key":"2008_CR53","unstructured":"Tu, S., Boczar, R., Simchowitz, M., Soltanolkotabi, M., Recht, B.: Low-rank solutions of linear matrix equations via procrustes flow. In: International Conference on Machine Learning, pp. 964\u2013973. PMLR (2016)"},{"issue":"3","key":"2008_CR54","doi-asserted-by":"publisher","first-page":"1198","DOI":"10.1137\/15M1050525","volume":"37","author":"K Wei","year":"2016","unstructured":"Wei, K., Cai, J.F., Chan, T.F., Leung, S.: Guarantees of riemannian optimization for low rank matrix recovery. SIAM J. Matrix Anal. Appl. 37(3), 1198\u20131222 (2016)","journal-title":"SIAM J. Matrix Anal. Appl."},{"key":"2008_CR55","doi-asserted-by":"crossref","unstructured":"Wei, K., Cai, J.F., Chan, T.F., Leung, S.: Guarantees of riemannian optimization for low rank matrix completion. Inverse Probl. Ima. 14(2) (2020)","DOI":"10.3934\/ipi.2020011"},{"key":"2008_CR56","unstructured":"Yalcin, B., Zhang, H., Lavaei, J., Sojoudi, S.: Factorization approach for low-complexity matrix completion problems: exponential number of spurious solutions and failure of gradient methods. In: International Conference on Artificial Intelligence and Statistics, pp. 1\u20139. PMLR (2022)"},{"key":"2008_CR57","unstructured":"Yi, X., Park, D., Chen, Y., Caramanis, C.: Fast algorithms for robust pca via gradient descent. Adv. Neural Inform. Process. Syst. 29 (2016)"},{"issue":"1","key":"2008_CR58","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1137\/19M1305045","volume":"3","author":"A Yurtsever","year":"2021","unstructured":"Yurtsever, A., Tropp, J.A., Fercoq, O., Udell, M., Cevher, V.: Scalable semidefinite programming. SIAM J. Math. Data Sci. 3(1), 171\u2013200 (2021)","journal-title":"SIAM J. Math. Data Sci."},{"key":"2008_CR59","unstructured":"Zhang, H., Bi, Y., Lavaei, J.: General low-rank matrix optimization: geometric analysis and sharper bounds. Adv. Neural Inform. Process. Syst. 34 (2021)"},{"key":"2008_CR60","doi-asserted-by":"crossref","unstructured":"Zhang, H., Yalcin, B., Lavaei, J., Sojoudi, S.: A new complexity metric for nonconvex rank-one generalized matrix completion. arXiv preprint arXiv:2204.02364 (2022)","DOI":"10.1007\/s10107-023-02008-5"},{"key":"2008_CR61","unstructured":"Zhang, J., Fattahi, S., Zhang, R.: Preconditioned gradient descent for over-parameterized nonconvex matrix factorization. Adv. Neural Inform. Process. Syst. 34 (2021)"},{"issue":"114","key":"2008_CR62","first-page":"1","volume":"20","author":"RY Zhang","year":"2019","unstructured":"Zhang, R.Y., Sojoudi, S., Lavaei, J.: Sharp restricted isometry bounds for the inexistence of spurious local minima in nonconvex matrix recovery. J. Mach. Learn. Res. 20(114), 1\u201334 (2019)","journal-title":"J. Mach. Learn. Res."},{"key":"2008_CR63","unstructured":"Zheng, Q., Lafferty, J.: A convergent gradient descent algorithm for rank minimization and semidefinite programming from random linear measurements. In: Proceedings of the 28th International Conference on Neural Information Processing Systems-Volume 1, pp. 109\u2013117 (2015)"},{"issue":"13","key":"2008_CR64","doi-asserted-by":"publisher","first-page":"3614","DOI":"10.1109\/TSP.2018.2835403","volume":"66","author":"Z Zhu","year":"2018","unstructured":"Zhu, Z., Li, Q., Tang, G., Wakin, M.B.: Global optimality in low-rank matrix optimization. IEEE Trans. Signal Process. 66(13), 3614\u20133628 (2018)","journal-title":"IEEE Trans. Signal Process."}],"container-title":["Mathematical Programming"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10107-023-02008-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10107-023-02008-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10107-023-02008-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,7]],"date-time":"2024-08-07T15:05:05Z","timestamp":1723043105000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10107-023-02008-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,16]]},"references-count":64,"journal-issue":{"issue":"1-2","published-print":{"date-parts":[[2024,9]]}},"alternative-id":["2008"],"URL":"https:\/\/doi.org\/10.1007\/s10107-023-02008-5","relation":{},"ISSN":["0025-5610","1436-4646"],"issn-type":[{"value":"0025-5610","type":"print"},{"value":"1436-4646","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,8,16]]},"assertion":[{"value":"5 April 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 July 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 August 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}