{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,27]],"date-time":"2025-11-27T18:13:22Z","timestamp":1764267202322,"version":"3.46.0"},"reference-count":62,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,3,28]],"date-time":"2024-03-28T00:00:00Z","timestamp":1711584000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,3,28]],"date-time":"2024-03-28T00:00:00Z","timestamp":1711584000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Mach Learn"],"published-print":{"date-parts":[[2024,5]]},"DOI":"10.1007\/s10994-024-06525-y","type":"journal-article","created":{"date-parts":[[2024,3,28]],"date-time":"2024-03-28T18:01:37Z","timestamp":1711648897000},"page":"3049-3071","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Personalization for web-based services using offline reinforcement learning"],"prefix":"10.1007","volume":"113","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8828-0793","authenticated-orcid":false,"given":"Pavlos Athanasios","family":"Apostolopoulos","sequence":"first","affiliation":[]},{"given":"Zehui","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Hanson","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Tenghyu","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Chad","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Kittipate","family":"Virochsiri","sequence":"additional","affiliation":[]},{"given":"Norm","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Igor L.","family":"Markov","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,28]]},"reference":[{"issue":"7","key":"6525_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3543846","volume":"55","author":"MM Afsar","year":"2022","unstructured":"Afsar, M. M., Crump, T., & Far, B. (2022). Reinforcement learning based recommender systems: A survey. ACM Computing Surveys, 55(7), 1\u201338.","journal-title":"ACM Computing Surveys"},{"issue":"1","key":"6525_CR2","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1177\/0278364919887447","volume":"39","author":"OM Andrychowicz","year":"2020","unstructured":"Andrychowicz, O. M., Baker, B., Chociej, M., Jozefowicz, R., McGrew, B., Pachocki, J., Petron, A., Plappert, M., Powell, G., Ray, A., et al. (2020). Learning dexterous in-hand manipulation. The International Journal of Robotics Research, 39(1), 3\u201320.","journal-title":"The International Journal of Robotics Research"},{"issue":"5","key":"6525_CR3","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1080\/07366981.2015.1104935","volume":"52","author":"A Awasthi","year":"2015","unstructured":"Awasthi, A. (2015). Reducing identity theft using one-time passwords and SMS. EDPACS, 52(5), 9\u201319.","journal-title":"EDPACS"},{"issue":"6624","key":"6525_CR4","doi-asserted-by":"publisher","first-page":"1067","DOI":"10.1126\/science.ade9097","volume":"378","author":"A Bakhtin","year":"2022","unstructured":"(FAIR)\u2020, M.F.A.R.D.T., Bakhtin, A., Brown, N., Dinan, E., Farina, G., Flaherty, C., Fried, D., Goff, A., Gray, J., Hu, H., et al. (2022). Human-level play in the game of diplomacy by combining language models with strategic reasoning. Science, 378(6624), 1067\u20131074.","journal-title":"Science"},{"key":"6525_CR5","unstructured":"Barth-Maron, G., Hoffman, M.W., Budden, D., Dabney, W., Horgan, D., Dhruva, T., Muldal, A., Heess, N., & Lillicrap, T. (2018). Distributed distributional deterministic policy gradients."},{"issue":"7836","key":"6525_CR6","doi-asserted-by":"publisher","first-page":"77","DOI":"10.1038\/s41586-020-2939-8","volume":"588","author":"MG Bellemare","year":"2020","unstructured":"Bellemare, M. G., Candido, S., Castro, P. S., Gong, J., Machado, M. C., Moitra, S., Ponda, S. S., & Wang, Z. (2020). Autonomous navigation of stratospheric balloons using reinforcement learning. Nature, 588(7836), 77\u201382.","journal-title":"Nature"},{"key":"6525_CR7","volume-title":"Dynamic Programming","author":"R Bellman","year":"1957","unstructured":"Bellman, R. (1957). Dynamic Programming (1st ed.). Princeton, USA: Princeton University Press.","edition":"1"},{"key":"6525_CR8","unstructured":"Brockman, G., Cheung, V., Pettersson, L., Schneider, J., Schulman, J., Tang, J., & Zaremba, W. (2016). OpenAI Gym."},{"key":"6525_CR9","unstructured":"Chen, X., Wang, S., McAuley, J., Jannach, D., & Yao, L. (2023). On the opportunities and challenges of offline reinforcement learning for recommender systems. arXiv preprint arXiv:2308.11336"},{"key":"6525_CR10","doi-asserted-by":"crossref","unstructured":"Dabney, W., Rowland, M., Bellemare, M.G., & Munos, R. (2018). Distributional reinforcement learning with quantile regression. In Thirty-Second AAAI Conference on Artificial Intelligence.","DOI":"10.1609\/aaai.v32i1.11791"},{"key":"6525_CR11","doi-asserted-by":"crossref","unstructured":"Deffayet, R., Thonet, T., Renders, J.-M., & de Rijke, M. (2023). Offline evaluation for reinforcement learning-based recommendation: A critical issue and some alternatives. In ACM SIGIR Forum, vol. 56, pp. 1\u201314. ACM New York, USA.","DOI":"10.1145\/3582900.3582905"},{"issue":"7897","key":"6525_CR12","doi-asserted-by":"publisher","first-page":"414","DOI":"10.1038\/s41586-021-04301-9","volume":"602","author":"J Degrave","year":"2022","unstructured":"Degrave, J., Felici, F., Buchli, J., Neunert, M., Tracey, B., Carpanese, F., Ewalds, T., Hafner, R., Abdolmaleki, A., de Las Casas, D., et al. (2022). Magnetic control of tokamak plasmas through deep reinforcement learning. Nature, 602(7897), 414\u2013419.","journal-title":"Nature"},{"key":"6525_CR13","unstructured":"Duncan, A.D. (2021). Over 100 data and analytics predictions through 2025."},{"key":"6525_CR14","unstructured":"Fujimoto, S., Meger, D., & Precup, D. (2019). Off-policy deep reinforcement learning without exploration. In International Conference on Machine Learning, pp. 2052\u20132062. PMLR."},{"key":"6525_CR15","unstructured":"Fujimoto, S., Meger, D., & Precup, D. (2019). Off-policy deep reinforcement learning without exploration."},{"key":"6525_CR16","unstructured":"Gauci, J., Conti, E., Liang, Y., Virochsiri, K., He, Y., Kaden, Z., Narayanan, V., & Ye, X. (2018). Horizon: Facebook\u2019s open source applied reinforcement learning platform. CoRR, 10 arXiv:1811.00260"},{"key":"6525_CR17","unstructured":"Hidasi, B., Karatzoglou, A., Baltrunas, L., & Tikk, D. (2016). Session-based recommendations with recurrent neural networks."},{"issue":"3","key":"6525_CR18","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1016\/j.eij.2015.06.005","volume":"16","author":"FO Isinkaye","year":"2015","unstructured":"Isinkaye, F. O., Folajimi, Y. O., & Ojokoh, B. A. (2015). Recommendation systems: Principles, methods and evaluation. Egyptian Informatics Journal, 16(3), 261\u2013273.","journal-title":"Egyptian Informatics Journal"},{"key":"6525_CR19","unstructured":"Jaques, N., Ghandeharioun, A., Shen, J.H., Ferguson, C., Lapedriza, A., Jones, N., Gu, S., & Picard, R. (2019). Way off-policy batch deep reinforcement learning of implicit human preferences in dialog."},{"key":"6525_CR20","unstructured":"Jiang, N., & Li, L. (2016). Doubly robust off-policy value evaluation for reinforcement learning. In Balcan, M.F., & Weinberger, K.Q. (eds.) Proceedings of The 33rd International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 48, pp. 652\u2013661. PMLR, New York, NYA. http:\/\/proceedings.mlr.press\/v48\/jiang16.html"},{"key":"6525_CR21","unstructured":"Kingma, D.P., & Ba, J. (2017). Adam: A method for stochastic optimization."},{"issue":"8","key":"6525_CR22","doi-asserted-by":"publisher","first-page":"922","DOI":"10.1007\/978-1-4899-7687-1_891","volume":"7","author":"R Kohavi","year":"2017","unstructured":"Kohavi, R., & Longbotham, R. (2017). Online controlled experiments and a\/b testing. Encyclopedia of Machine Learning and Data Mining, 7(8), 922\u2013929.","journal-title":"Encyclopedia of Machine Learning and Data Mining"},{"key":"6525_CR23","unstructured":"Konda, V., & Tsitsiklis, J. (2000). Actor-critic algorithms. In: Solla, S., Leen, T., & M\u00fcller, K. (eds.) Advances in Neural Information Processing Systems, vol. 12, pp. 1008\u20131014. MIT Press, Denver. https:\/\/proceedings.neurips.cc\/paper\/1999\/file\/6449f44a102fde848669bdd9eb6b76fa-Paper.pdf"},{"key":"6525_CR24","unstructured":"Kumar, A., Fu, J., Soh, M., Tucker, G., & Levine, S. (2019). Stabilizing off-policy q-learning via bootstrapping error reduction. In Wallach, H., Larochelle, H., Beygelzimer, A., d\u2019 Alch\u00e9-Buc, F., Fox, E., & Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 32, pp. 11784\u201311794. Curran Associates, Inc., Virtual. https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/c2073ffa77b5357a498057413bb09d3a-Paper.pdf"},{"issue":"10","key":"6525_CR25","doi-asserted-by":"publisher","first-page":"4156","DOI":"10.1073\/pnas.1804597116","volume":"116","author":"SR K\u00fcnzel","year":"2019","unstructured":"K\u00fcnzel, S. R., Sekhon, J. S., Bickel, P. J., & Yu, B. (2019). Metalearners for estimating heterogeneous treatment effects using machine learning. Proceedings of the National Academy of Sciences, 116(10), 4156\u20134165.","journal-title":"Proceedings of the National Academy of Sciences"},{"key":"6525_CR26","unstructured":"Ledesma, R.M. (2020). Systems and methods for one-time password authentication. Google Patents. US Patent App. 16\/918,742."},{"key":"6525_CR27","unstructured":"Levine, S., Kumar, A., Tucker, G., & Fu, J. (2020). Offline Reinforcement Learning: Tutorial, review, and perspectives on open problems."},{"key":"6525_CR28","doi-asserted-by":"publisher","unstructured":"Li, L., Chu, W., Langford, J., & Schapire, R.E. (2010). A contextual-bandit approach to personalized news article recommendation. Proceedings of the 19th international conference on World wide web - WWW \u201910. https:\/\/doi.org\/10.1145\/1772690.1772758","DOI":"10.1145\/1772690.1772758"},{"key":"6525_CR29","doi-asserted-by":"publisher","unstructured":"Li, J., Ren, P., Chen, Z., Ren, Z., Lian, T., & Ma, J. (2017). Neural attentive session-based recommendation. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management. CIKM \u201917, pp. 1419\u20131428. ACM, New York. https:\/\/doi.org\/10.1145\/3132847.3132926","DOI":"10.1145\/3132847.3132926"},{"key":"6525_CR30","unstructured":"Lillicrap, T.P., Hunt, J.J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., Silver, D., & Wierstra, D. (2019). Continuous control with deep reinforcement learning."},{"key":"6525_CR31","unstructured":"Liu, Y., Swaminathan, A., Agarwal, A., & Brunskill, E. (2020). Provably good batch reinforcement learning without great exploration."},{"key":"6525_CR32","doi-asserted-by":"crossref","unstructured":"Ma, J., Zhao, Z., Yi, X., Yang, J., Chen, M., Tang, J., Hong, L., & Chi, E.H. (2020). Off-policy learning in two-stage recommender systems. In Proceedings of The Web Conference 2020, pp. 463\u2013473.","DOI":"10.1145\/3366423.3380130"},{"key":"6525_CR33","doi-asserted-by":"crossref","unstructured":"Markov, I.L., Wang, H., Kasturi, N., Singh, S., Yuen, S.W., Garrard, M., Tran, S., Huang, Y., Wang, Z., Glotov, I., Gupta, T., Huang, B., Chen, P., Xie, X., Belkin, M., Uryasev, S., Howie, S., Bakshy, E., & Zhou, N. (2022). Looper: An end-to-end ML platform for product decisions.","DOI":"10.1145\/3534678.3539059"},{"key":"6525_CR34","unstructured":"Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., & Riedmiller, M. (2013). Playing Atari with deep reinforcement learning."},{"key":"6525_CR35","doi-asserted-by":"crossref","unstructured":"Nambiar, M., Ghosh, S., Ong, P., Chan, Y.E., Bee, Y.M., & Krishnaswamy, P. (2023). Deep offline reinforcement learning for real-world treatment optimization applications. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 4673\u20134684.","DOI":"10.1145\/3580305.3599800"},{"key":"6525_CR36","unstructured":"Ng, A.Y., & Russell, S.J. (2000). Algorithms for inverse reinforcement learning. In Proceedings of the Seventeenth International Conference on Machine Learning. ICML \u201900, pp. 663\u2013670. Morgan Kaufmann Publishers Inc., San Francisco, USA."},{"key":"6525_CR37","doi-asserted-by":"publisher","unstructured":"Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C.L., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., Schulman, J., Hilton, J., Kelton, F., Miller, L., Simens, M., Askell, A., Welinder, P., Christiano, P., Leike, J., & Lowe, R. (2022). Training language models to follow instructions with human feedback. arXiv. https:\/\/doi.org\/10.48550\/ARXIV.2203.02155. arXiv:2203.02155","DOI":"10.48550\/ARXIV.2203.02155"},{"key":"6525_CR38","unstructured":"Paine, T.L., Paduraru, C., Michi, A., Gulcehre, C., Zolna, K., Novikov, A., Wang, Z., & de Freitas, N. (2020). Hyperparameter selection for offline reinforcement learning."},{"key":"6525_CR39","doi-asserted-by":"crossref","unstructured":"Pham, M., Nguyen, H., Dang, L., & Nieves, J.A. (2021). Compressive features in offline reinforcement learning for recommender systems. In 2021 IEEE International Conference on Big Data (Big Data), pp. 5719\u20135726. IEEE.","DOI":"10.1109\/BigData52589.2021.9671419"},{"key":"6525_CR40","doi-asserted-by":"publisher","unstructured":"Pradel, B., Sean, S., Delporte, J., Gu\u00e9rif, S., Rouveirol, C., Usunier, N., Fogelman-Souli\u00e9, F., & Dufau-Joel, F. (2011). A case study in a recommender system based on purchase data. In Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. KDD \u201911, pp. 377\u2013385. ACM, New York. https:\/\/doi.org\/10.1145\/2020408.2020470","DOI":"10.1145\/2020408.2020470"},{"key":"6525_CR41","unstructured":"Precup, D., Sutton, R.S., & Singh, S.P. (2000). Eligibility traces for off-policy policy evaluation. In Proceedings of the Seventeenth International Conference on Machine Learning. ICML \u201900, pp. 759\u2013766. Morgan Kaufmann Publishers Inc., San Francisco, USA."},{"key":"6525_CR42","volume-title":"Markov Decision Processes: Discrete Stochastic Dynamic Programming","author":"ML Puterman","year":"2014","unstructured":"Puterman, M. L. (2014). Markov Decision Processes: Discrete Stochastic Dynamic Programming. Virtual: John Wiley & Sons."},{"key":"6525_CR43","unstructured":"Rockwell, P. (2016). Two factor authentication using a one-time password. Google Patents. US Patent 9,378,356."},{"key":"6525_CR44","doi-asserted-by":"crossref","unstructured":"Saito, Y., & Joachims, T. (2021). Counterfactual learning and evaluation for recommender systems: Foundations, implementations, and recent advances. In Proceedings of the 15th ACM Conference on Recommender Systems, pp. 828\u2013830.","DOI":"10.1145\/3460231.3473320"},{"key":"6525_CR45","unstructured":"Schulman, J., Levine, S., Abbeel, P., Jordan, M., & Moritz, P. (2015). Trust region policy optimization. In Bach, F., & Blei, D. (eds.) Proceedings of the 32nd International Conference on Machine Learning. Proc. Machine Learning Research, vol. 37, pp. 1889\u20131897. PMLR, Lille, France. http:\/\/proceedings.mlr.press\/v37\/schulman15.html"},{"issue":"3","key":"6525_CR46","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1002\/spy2.102","volume":"3","author":"MK Sharma","year":"2020","unstructured":"Sharma, M. K., & Nene, M. J. (2020). Two-factor authentication using biometric based quantum operations. Security and Privacy, 3(3), 102.","journal-title":"Security and Privacy"},{"key":"6525_CR47","unstructured":"Siegel, N.Y., Springenberg, J.T., Berkenkamp, F., Abdolmaleki, A., Neunert, M., Lampe, T., Hafner, R., Heess, N., & Riedmiller, M. (2020). Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning."},{"key":"6525_CR48","unstructured":"Sutton, R.S., McAllester, D., Singh, S., & Mansour, Y. (1999). Policy gradient methods for reinforcement learning with function approximation. In Proceedings of the 12th International Conference on Neural Information Processing Systems. NIPS\u201999, pp. 1057\u20131063. MIT Press, Cambridge, USA."},{"key":"6525_CR49","volume-title":"Reinforcement Learning: An Introduction","author":"RS Sutton","year":"2018","unstructured":"Sutton, R. S., & Barto, A. G. (2018). Reinforcement Learning: An Introduction. Cambridge: A Bradford Book."},{"issue":"1","key":"6525_CR50","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1111\/bmsp.12144","volume":"72","author":"X Tang","year":"2019","unstructured":"Tang, X., Chen, Y., Li, X., Liu, J., & Ying, Z. (2019). A reinforcement learning approach to personalized learning recommendation systems. British Journal of Mathematical and Statistical Psychology, 72(1), 108\u2013135.","journal-title":"British Journal of Mathematical and Statistical Psychology"},{"key":"6525_CR51","unstructured":"Thomas, P., & Brunskill, E. (2016). Data-efficient off-policy policy evaluation for reinforcement learning. In Balcan, M.F., & Weinberger, K.Q. (eds.) Proceedings of The 33rd International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 48, pp. 2139\u20132148. PMLR, New York. http:\/\/proceedings.mlr.press\/v48\/thomasa16.html"},{"key":"6525_CR52","doi-asserted-by":"crossref","unstructured":"van Hasselt, H., Guez, A., & Silver, D. (2015). Deep reinforcement learning with double Q-learning.","DOI":"10.1609\/aaai.v30i1.10295"},{"key":"6525_CR53","doi-asserted-by":"publisher","unstructured":"Wang, Z., Jiang, Z., Ren, Z., Tang, J., & Yin, D. (2018). A path-constrained framework for discriminating substitutable and complementary products in e-commerce. In Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining. WSDM \u201918, pp. 619\u2013627. ACM, New York. https:\/\/doi.org\/10.1145\/3159652.3159710","DOI":"10.1145\/3159652.3159710"},{"key":"6525_CR54","unstructured":"Wang, Z., Novikov, A., Zolna, K., Springenberg, J.T., Reed, S., Shahriari, B., Siegel, N., Merel, J., Gulcehre, C., Heess, N., & de Freitas, N. (2020). Critic regularized regression."},{"key":"6525_CR55","unstructured":"Wang, Z., Schaul, T., Hessel, M., Hasselt, H., Lanctot, M., & Freitas, N. (2016). Dueling network architectures for deep reinforcement learning. In Balcan, M.F., & Weinberger, K.Q. (eds.) Proceedings of the 33rd International Conference on Machine Learning. Proc. Machine Learning Research, vol. 48, pp. 1995\u20132003. PMLR, New York, NY. http:\/\/proceedings.mlr.press\/v48\/wangf16.html"},{"key":"6525_CR56","unstructured":"Wang, H., Wang, Z., & Ma, Y. (2020). Predictive precompute with recurrent neural networks. In Dhillon, I., Papailiopoulos, D., & Sze, V. (eds.) Proceedings of the Machine Learning and Systems, vol. 2, pp. 470\u2013480. mlsys.org, San Jose. https:\/\/proceedings.mlsys.org\/paper\/2020\/file\/8f53295a73878494e9bc8dd6c3c7104f-Paper.pdf"},{"issue":"3\u20134","key":"6525_CR57","doi-asserted-by":"publisher","first-page":"279","DOI":"10.1007\/BF00992698","volume":"8","author":"CJ Watkins","year":"1992","unstructured":"Watkins, C. J., & Dayan, P. (1992). Q-learning. Machine Learning, 8(3\u20134), 279\u2013292.","journal-title":"Machine Learning"},{"key":"6525_CR58","unstructured":"Wu, Y., Tucker, G., & Nachum, O. (2019). Behavior regularized offline reinforcement learning. arXiv preprint arXiv:1911.11361"},{"key":"6525_CR59","doi-asserted-by":"crossref","unstructured":"Xiao, T., & Wang, D. (2021). A general offline reinforcement learning framework for interactive recommendation. In Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 4512\u20134520.","DOI":"10.1609\/aaai.v35i5.16579"},{"key":"6525_CR60","doi-asserted-by":"crossref","unstructured":"Xin, X., Pimentel, T., Karatzoglou, A., Ren, P., Christakopoulou, K., & Ren, Z. (2022). Rethinking reinforcement learning for recommendation: A prompt perspective. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1347\u20131357.","DOI":"10.1145\/3477495.3531714"},{"key":"6525_CR61","doi-asserted-by":"crossref","unstructured":"Yuan, Y., Muralidharan, A., Nandy, P., Cheng, M., & Prabhakar, P. (2022). Offline reinforcement learning for mobile notifications. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, pp. 3614\u20133623.","DOI":"10.1145\/3511808.3557083"},{"key":"6525_CR62","unstructured":"Ziebart, B.D., Maas, A.L., Bagnell, J.A., Dey, A.K., et al. (2008). Maximum entropy inverse reinforcement learning. In Aaai, vol. 8, pp. 1433\u20131438. Chicago, USA."}],"container-title":["Machine Learning"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10994-024-06525-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10994-024-06525-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10994-024-06525-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,27]],"date-time":"2025-11-27T18:08:36Z","timestamp":1764266916000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10994-024-06525-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3,28]]},"references-count":62,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,5]]}},"alternative-id":["6525"],"URL":"https:\/\/doi.org\/10.1007\/s10994-024-06525-y","relation":{},"ISSN":["0885-6125","1573-0565"],"issn-type":[{"type":"print","value":"0885-6125"},{"type":"electronic","value":"1573-0565"}],"subject":[],"published":{"date-parts":[[2024,3,28]]},"assertion":[{"value":"9 February 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 October 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 February 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 March 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"Not applicable.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"This content has been made available to all.","name":"free","label":"Free to read"}]}}