{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,15]],"date-time":"2026-01-15T08:47:30Z","timestamp":1768466850018,"version":"3.49.0"},"reference-count":69,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2025,2,28]],"date-time":"2025-02-28T00:00:00Z","timestamp":1740700800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,2,28]],"date-time":"2025-02-28T00:00:00Z","timestamp":1740700800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["72293573"],"award-info":[{"award-number":["72293573"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["72293575"],"award-info":[{"award-number":["72293575"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Strategic Priority Research Program of Chinese Academy of Sciences","award":["XDA27030100"],"award-info":[{"award-number":["XDA27030100"]}]},{"name":"Excellent Youth Program of State Key Laboratory of Multimodal Artificial Intelligence Systems","award":["MAIS2024310"],"award-info":[{"award-number":["MAIS2024310"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2025,4]]},"DOI":"10.1007\/s40747-025-01785-0","type":"journal-article","created":{"date-parts":[[2025,2,28]],"date-time":"2025-02-28T05:35:55Z","timestamp":1740720955000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Demonstration and offset augmented meta reinforcement learning with sparse rewards"],"prefix":"10.1007","volume":"11","author":[{"given":"Haorui","family":"Li","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4977-4093","authenticated-orcid":false,"given":"Jiaqi","family":"Liang","sequence":"additional","affiliation":[]},{"given":"Xiaoxuan","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Chengzhi","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Linjing","family":"Li","sequence":"additional","affiliation":[]},{"given":"Daniel","family":"Zeng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,28]]},"reference":[{"key":"1785_CR1","unstructured":"Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of the International Conference on Machine Learning, 70:1126\u20131135"},{"key":"1785_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.inffus.2022.03.003","volume":"85","author":"P Ladosz","year":"2022","unstructured":"Ladosz P, Weng L, Kim M, Oh H (2022) Exploration in deep reinforcement learning: a survey. Inf Fus 85:1\u201322","journal-title":"Inf Fus"},{"key":"1785_CR3","doi-asserted-by":"crossref","unstructured":"Hong W, Zhu M, Liu M, Zhang W, Zhou M, Yu Y, Sun P (2019) Generative adversarial exploration for reinforcement learning. In: Proceedings of the International Conference on Distributed Artificial Intelligence, pp. 1\u201310","DOI":"10.1145\/3356464.3357706"},{"issue":"11","key":"1785_CR4","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2020) Generative adversarial networks. Commun ACM 63(11):139\u2013144","journal-title":"Commun ACM"},{"key":"1785_CR5","unstructured":"Stadie BC, Levine S, Abbeel P (2015) Incentivizing exploration in reinforcement learning with deep predictive models. arXiv preprint arXiv:1507.00814"},{"issue":"2","key":"1785_CR6","doi-asserted-by":"publisher","first-page":"1086","DOI":"10.1007\/s10489-020-01849-3","volume":"51","author":"N Bougie","year":"2021","unstructured":"Bougie N, Ichise R (2021) Fast and slow curiosity for high-level exploration in reinforcement learning. Appl Intell 51(2):1086\u20131107","journal-title":"Appl Intell"},{"key":"1785_CR7","doi-asserted-by":"crossref","unstructured":"Bougie N, Ichise R (2020) Towards high-level intrinsic exploration in reinforcement learning. In: Proceedings of the International Joint Conference on Artificial Intelligence, IJCAI 2020, pp. 5186\u20135187","DOI":"10.24963\/ijcai.2020\/733"},{"key":"1785_CR8","unstructured":"Burda Y, Edwards H, Storkey AJ, Klimov O (2019) Exploration by random network distillation. In: Proceedings of the International Conference on Learning Representations"},{"key":"1785_CR9","doi-asserted-by":"crossref","unstructured":"Kolter JZ, Ng AY (2009) Near-bayesian exploration in polynomial time. In: Proceedings of the International Conference on Machine Learning 382:513\u2013520","DOI":"10.1145\/1553374.1553441"},{"key":"1785_CR10","unstructured":"Pathak D, Gandhi D, Gupta A (2019) Self-supervised exploration via disagreement. In: Proceedings of TheInternational Conference on Machine Learning 97:5062\u20135071"},{"key":"1785_CR11","unstructured":"Houthooft R, Chen X, Duan Y, Schulman J, Turck FD, Abbeel P (2016) VIME: variational information maximizing exploration. In: Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 1109\u20131117"},{"issue":"2","key":"1785_CR12","doi-asserted-by":"publisher","first-page":"189","DOI":"10.1162\/EVCO_a_00025","volume":"19","author":"J Lehman","year":"2011","unstructured":"Lehman J, Stanley KO (2011) Abandoning objectives: evolution through the search for novelty alone. Evol Comput 19(2):189\u2013223","journal-title":"Evol Comput"},{"key":"1785_CR13","unstructured":"Eysenbach B, Gupta A, Ibarz J, Levine S (2019) Diversity is all you need: learning skills without a reward function. In: Proceedings of the International Conference on Learning Representations"},{"key":"1785_CR14","unstructured":"Gangwani T, Liu Q, Peng J (2019) Learning self-imitating diverse policies. In: Proceedings of the International Conference on Learning Representations"},{"key":"1785_CR15","unstructured":"Pong VH, Nair A, Smith LM, Huang C, Levine S (2022) Offline meta-reinforcement learning with online self-supervision. In: Proceedings of the International Conference on Machine Learning 162:17811\u201317829"},{"key":"1785_CR16","unstructured":"Packer C, Abbeel P, Gonzalez JE (2021) Hindsight task relabelling: experience replay for sparse reward meta-RL. In: Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 2466\u20132477"},{"key":"1785_CR17","doi-asserted-by":"crossref","unstructured":"Guo Y, Wu Q, Lee H (2022) Learning action translator for meta reinforcement learning on sparse-reward tasks. In: Proceeding of the AAAI Conference on Artificial Intelligence, pp. 6792\u20136800","DOI":"10.1609\/aaai.v36i6.20635"},{"key":"1785_CR18","doi-asserted-by":"crossref","unstructured":"Hua Y, Wang X, Jin B, Li W, Yan J, He X, Zha H (2021) HMRL: hyper-meta learning for sparse reward reinforcement learning problem. In: Proceeding of the ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 637\u2013645","DOI":"10.1145\/3447548.3467242"},{"key":"1785_CR19","unstructured":"Nam T, Sun S-H, Pertsch K, Hwang SJ, Lim JJ (2022) Skill-based meta-reinforcement learning. In: Proceedings of the International Conference on Learning Representations"},{"key":"1785_CR20","unstructured":"Zhang J, Wang J, Hu H, Chen T, Chen Y, Fan C, Zhang C (2021) MetaCURE: meta reinforcement learning with empowerment-driven exploration. In: Proceedings of the International Conference on Machine Learning, vol. 139, pp. 12600\u201312610"},{"key":"1785_CR21","unstructured":"Zintgraf LM, Feng L, Lu C, Igl M, Hartikainen K, Hofmann K, Whiteson S (2021) Exploration in approximate hyper-state space for meta reinforcement learning. In: Proceedings of the International Conference on Machine Learning 139:12991\u201313001"},{"key":"1785_CR22","unstructured":"Vecerik M, Hester T, Scholz J, Wang F, Pietquin O, Piot B, Heess N, Roth\u00f6rl T, Lampe T, Riedmiller M (2017) Leveraging demonstrations for deep reinforcement learning on robotics problems with sparse rewards. arXiv preprint arXiv:1707.08817"},{"key":"1785_CR23","doi-asserted-by":"crossref","unstructured":"Hester T, Vecer\u00edk M, Pietquin O, Lanctot M, Schaul T, Piot B, Horgan D, Quan J, Sendonaris A, Osband I, Dulac-Arnold G, Agapiou JP, Leibo JZ, Gruslys A (2018) Deep Q-learning From Demonstrations. In: Proceeding of the AAAI Conference on Artificial Intelligence, pp. 3223\u20133230","DOI":"10.1609\/aaai.v32i1.11757"},{"key":"1785_CR24","doi-asserted-by":"crossref","unstructured":"Nair A, McGrew B, Andrychowicz M, Zaremba W, Abbeel P (2018) Overcoming exploration in reinforcement learning with demonstrations. In: Proceedings of the IEEE International Conference on Robotics and Automation, pp. 6292\u20136299","DOI":"10.1109\/ICRA.2018.8463162"},{"key":"1785_CR25","doi-asserted-by":"crossref","unstructured":"Rajeswaran A, Kumar V, Gupta A, Vezzani G, Schulman J, Todorov E, Levine S (2018) Learning complex dexterous manipulation with deep reinforcement learning and demonstrations. In: Proceeding of the Robotics: Science and Systems","DOI":"10.15607\/RSS.2018.XIV.049"},{"key":"1785_CR26","unstructured":"Rengarajan D, Vaidya G, Sarvesh A, Kalathil DM, Shakkottai S (2022) Reinforcement learning with sparse rewards using guidance from offline demonstration. In: Proceeding of the International Conference on Learning Representations"},{"key":"1785_CR27","unstructured":"Finn C, Yu T, Zhang T, Abbeel P, Levine S (2017) One-shot visual imitation learning via meta-learning. In: Proceeding of the Annual Conference on Robot Learning, 78:357\u2013368"},{"key":"1785_CR28","unstructured":"Zhou A, Jang E, Kappler D, Herzog A, Khansari M, Wohlhart P, Bai Y, Kalakrishnan M, Levine S, Finn C (2020) Watch, try, learn: meta-learning from demonstrations and rewards. In: Proceeding of the International Conference on Learning Representations"},{"key":"1785_CR29","unstructured":"Mendonca R, Gupta A, Kralev R, Abbeel P, Levine S, Finn C (2019) Guided meta-policy search. In: Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 9653\u20139664"},{"key":"1785_CR30","unstructured":"Rengarajan D, Chaudhary S, Kim J, Kalathil D, Shakkottai S (2022) Enhanced meta reinforcement learning via demonstrations in sparse reward environments. In: Proceedings of the Annual Conference on Neural Information Processing Systems"},{"key":"1785_CR31","doi-asserted-by":"crossref","unstructured":"Peng Z, Han C, Liu Y, Zhou Z (2023) Weighted policy constraints for offline reinforcement learning. In: Proceedings of the AAAI Conference on Artificial Intelligence 37:9435\u20139443","DOI":"10.1609\/aaai.v37i8.26130"},{"key":"1785_CR32","unstructured":"Yang Y, Caluwaerts K, Iscen A, Tan J, Finn C (2019) NoRML: no-reward meta learning. In: Proceedings of the International Conference on Autonomous Agents and MultiAgent Systems, pp. 323\u2013331"},{"key":"1785_CR33","unstructured":"Raghu A, Raghu M, Bengio S, Vinyals O (2020) Rapid learning or feature reuse? Towards understanding the effectiveness of MAML. In: Proceeding of the International Conference on Learning Representations"},{"key":"1785_CR34","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Identity mappings in deep residual networks. In: Proceedings of the European Conference on Computer Vision, vol. 9908, pp. 630\u2013645","DOI":"10.1007\/978-3-319-46493-0_38"},{"key":"1785_CR35","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2021) An image is worth 16x16 words: transformers for image recognition at scale. In: Proceeding of the International Conference on Learning Representations"},{"key":"1785_CR36","doi-asserted-by":"crossref","unstructured":"Li Z, Tucker R, Snavely N, Holynski A (2024) Generative image dynamics. In: Proceeding of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 24142\u201324153","DOI":"10.1109\/CVPR52733.2024.02279"},{"key":"1785_CR37","unstructured":"Rafailov R, Sharma A, Mitchell E, Manning CD, Ermon S, Finn C (2023) Direct preference optimization: your language model is secretly a reward model. In: Proceedings of the Annual Conference on Neural Information Processing Systems"},{"issue":"4","key":"1785_CR38","doi-asserted-by":"publisher","first-page":"1445","DOI":"10.1109\/TCDS.2024.3370219","volume":"16","author":"Z Xiao","year":"2024","unstructured":"Xiao Z, Xu X, Xing H, Zhao B, Wang X, Song F, Qu R, Feng L (2024) DTCM: deep transformer capsule mutual distillation for multivariate time series classification. IEEE Trans Cogn Dev Syst 16(4):1445\u20131461","journal-title":"IEEE Trans Cogn Dev Syst"},{"key":"1785_CR39","unstructured":"Xiao Z, Tong H, Qu R, Xing H, Luo S, Zhu Z, Song F, Feng L (2023) CapMatch: semi-supervised contrastive transformer capsule with feature-based knowledge distillation for human activity recognition. IEEE Transactions on Neural Networks and Learning Systems, 1\u201315"},{"key":"1785_CR40","unstructured":"Koch G, Zemel R, Salakhutdinov R, et al (2015) Siamese neural networks for one-shot image recognition. In: Proceedings of the International Conference on Machine Learning on Deep Learning Workshop, 2:1\u201330"},{"key":"1785_CR41","unstructured":"Vinyals O, Blundell C, Lillicrap T, Kavukcuoglu K, Wierstra D (2016) Matching networks for one shot learning. In: Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 3630\u20133638"},{"key":"1785_CR42","unstructured":"Snell J, Swersky K, Zemel RS (2017) Prototypical networks for few-shot learning. In: Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 4077\u20134087"},{"key":"1785_CR43","unstructured":"Nichol A, Schulman J (2018) Reptile: a scalable metalearning algorithm. arXiv preprint arXiv:1803.02999"},{"key":"1785_CR44","doi-asserted-by":"crossref","unstructured":"Harrison J, Sharma A, Pavone M (2018) Meta-learning priors for efficient online bayesian regression. In: Proceedings of the Workshop on the Algorithmic Foundations of Robotics 14:318\u2013337","DOI":"10.1007\/978-3-030-44051-0_19"},{"key":"1785_CR45","unstructured":"Duan Y, Schulman J, Chen X, Bartlett PL, Sutskever I, Abbeel P (2016) RL2: fast reinforcement learning via slow reinforcement learning. arXiv preprint arXiv:1611.02779"},{"key":"1785_CR46","unstructured":"Mishra N, Rohaninejad M, Chen X, Abbeel P (2018) A simple neural attentive meta-learner. In: Proceeding of the International Conference on Learning Representations"},{"key":"1785_CR47","unstructured":"Stadie BC, Yang G, Houthooft R, Chen X, Duan Y, Wu Y, Abbeel P, Sutskever I (2018) Some considerations on learning to explore via meta-reinforcement learning. arXiv preprint arXiv:1803.01118"},{"key":"1785_CR48","unstructured":"Wang JX, Kurth-Nelson Z, Tirumala D, Soyer H, Leibo JZ, Munos R, Blundell C, Kumaran D, Botvinick M (2016) Learning to reinforcement learn. arXiv preprint arXiv:1611.05763"},{"key":"1785_CR49","unstructured":"Gupta A, Mendonca R, Liu Y, Abbeel P, Levine S (2018) Meta-reinforcement learning of structured exploration strategies. In: Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 5307\u20135316"},{"key":"1785_CR50","unstructured":"Houthooft R, Chen Y, Isola P, Stadie BC, Wolski F, Ho J, Abbeel P (2018) Evolved policy gradients. In: Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 5405\u20135414"},{"key":"1785_CR51","unstructured":"Sung F, Zhang L, Xiang T, Hospedales T, Yang Y (2017) Learning to learn: meta-critic networks for sample efficient learning. arXiv preprint arXiv:1706.09529"},{"key":"1785_CR52","unstructured":"Beck J, Vuorio R, Liu EZ, Xiong Z, Zintgraf L, Finn C, Whiteson S (2023) A survey of meta-reinforcement learning. arXiv preprint arXiv:2301.08028"},{"key":"1785_CR53","unstructured":"Rothfuss J, Lee D, Clavera I, Asfour T, Abbeel P (2019) ProMP: proximal meta-policy search. In: Proceeding of the International Conference on Learning Representations"},{"key":"1785_CR54","unstructured":"Foerster JN, Farquhar G, Al-Shedivat M, Rockt\u00e4schel T, Xing EP, Whiteson S (2018) DiCE: the infinitely differentiable monte carlo estimator. In: Proceedings of the International Conference on Machine Learning 80:1524\u20131533"},{"key":"1785_CR55","unstructured":"Al-Shedivat M, Bansal T, Burda Y, Sutskever I, Mordatch I, Abbeel P (2018) Continuous adaptation via meta-learning in nonstationary and competitive environments. In: Proceeding of the International Conference on Learning Representations"},{"key":"1785_CR56","unstructured":"Packer C, Abbeel P, Gonzalez JE (2021) Hindsight task relabelling: experience replay for sparse reward Meta-RL. In: Ranzato M, Beygelzimer A, Dauphin YN, Liang P, Vaughan JW (eds) Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 2466\u20132477"},{"key":"1785_CR57","unstructured":"Xue Z, Peng Z, Li Q, Liu Z, Zhou B (2023) Guarded policy optimization with imperfect online demonstrations. In: Proceeding of the International Conference on Learning Representations"},{"key":"1785_CR58","unstructured":"Sutton RS, Barto AG (2018) Reinforcement learning: an introduction"},{"key":"1785_CR59","unstructured":"Sutton RS, McAllester DA, Singh S, Mansour Y (1999) Policy gradient methods for reinforcement learning with function approximation. In: Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 1057\u20131063"},{"key":"1785_CR60","unstructured":"Schulman J, Moritz P, Levine S, Jordan MI, Abbeel P (2016) High-dimensional continuous control using generalized advantage estimation. In: Proceeding of the International Conference on Learning Representations"},{"key":"1785_CR61","unstructured":"Haarnoja T, Zhou A, Abbeel P, Levine S (2018) Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In: Proceedings of the International Conference on Machine Learning 80:1856\u20131865"},{"key":"1785_CR62","unstructured":"Schulman J, Levine S, Abbeel P, Jordan MI, Moritz P (2015) Trust region policy optimization. In: Proceedings of the International Conference on Machine Learning 37:1889\u20131897"},{"key":"1785_CR63","unstructured":"Kingma DP, Ba J (2015) Adam: a method for stochastic optimization. In: Proceedings of the International Conference on Learning Representations"},{"key":"1785_CR64","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Identity mappings in deep residual networks. In: Proceedings of the European Conference on Computer Vision 9908:630\u2013645","DOI":"10.1007\/978-3-319-46493-0_38"},{"key":"1785_CR65","doi-asserted-by":"crossref","unstructured":"Todorov E, Erez T, Tassa Y (2012) Mujoco: a physics engine for model-based control. In: Proceedings of the IEEE\/RSJ International Conference on Intelligent Robots and Systems, pp. 5026\u20135033","DOI":"10.1109\/IROS.2012.6386109"},{"key":"1785_CR66","unstructured":"Rakelly K, Zhou A, Finn C, Levine S, Quillen D (2019) Efficient off-policy meta-reinforcement learning via probabilistic context variables. In: Proceedings of the International Conference on Machine Learning 97:5331\u20135340"},{"key":"1785_CR67","doi-asserted-by":"crossref","unstructured":"Zhou R, Gao C, Zhang Z, Yu Y (2024) Generalizable task representation learning for offline meta-reinforcement learning with data limitations. In: Proceeding of the AAAI Conference on Artificial Intelligence, pp. 17132\u201317140","DOI":"10.1609\/aaai.v38i15.29658"},{"key":"1785_CR68","unstructured":"Gao Y, Zhang R, Guo J, Wu F, Yi Q, Peng S, Lan S, Chen R, Du Z, Hu X, Guo Q, Li L, Chen Y (2023) Context shift reduction for offline meta-reinforcement learning. In: Proceedings of the Annual Conference on Neural Information Processing Systems, pp. 80024\u201380043"},{"key":"1785_CR69","doi-asserted-by":"publisher","first-page":"229","DOI":"10.1007\/BF00992696","volume":"8","author":"RJ Williams","year":"1992","unstructured":"Williams RJ (1992) Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach Learn 8:229\u2013256","journal-title":"Mach Learn"}],"updated-by":[{"DOI":"10.1007\/s40747-025-01933-6","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2025,5,19]],"date-time":"2025-05-19T00:00:00Z","timestamp":1747612800000}}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01785-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-025-01785-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01785-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,20]],"date-time":"2025-05-20T13:18:19Z","timestamp":1747747099000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-025-01785-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,28]]},"references-count":69,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,4]]}},"alternative-id":["1785"],"URL":"https:\/\/doi.org\/10.1007\/s40747-025-01785-0","relation":{"correction":[{"id-type":"doi","id":"10.1007\/s40747-025-01933-6","asserted-by":"object"}]},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,2,28]]},"assertion":[{"value":"5 August 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 January 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 February 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 May 2025","order":4,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Correction","order":5,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"A Correction to this paper has been published:","order":6,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"https:\/\/doi.org\/10.1007\/s40747-025-01933-6","URL":"https:\/\/doi.org\/10.1007\/s40747-025-01933-6","order":7,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"194"}}