{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T15:37:38Z","timestamp":1774021058822,"version":"3.50.1"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"4-5","license":[{"start":{"date-parts":[[2024,6,4]],"date-time":"2024-06-04T00:00:00Z","timestamp":1717459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,6,4]],"date-time":"2024-06-04T00:00:00Z","timestamp":1717459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"Honda Research Institute, EU, Europe"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Auton Robot"],"published-print":{"date-parts":[[2024,7]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>The advent of collaborative robots allows humans and robots to cooperate in a direct and physical way. While this leads to amazing new opportunities to create novel robotics applications, it is challenging to make the collaboration intuitive for the human. From a system\u2019s perspective, understanding the human intentions seems to be one promising way to get there. However, human behavior exhibits large variations between individuals, such as for instance preferences or physical abilities. This paper presents a novel concept for simultaneously learning a model of the human intentions and preferences incrementally during collaboration with a robot. Starting out with a nominal model, the system acquires collaborative skills step-by-step within only very few trials. The concept is based on a combination of model-based reinforcement learning and inverse reinforcement learning, adapted to fit collaborations in which human and robot think and act independently. We test the method and compare it to two baselines: one that imitates the human and one that uses plain maximum entropy inverse reinforcement learning, both in simulation and in a user study with a Franka Emika Panda robot arm.<\/jats:p>","DOI":"10.1007\/s10514-024-10167-3","type":"journal-article","created":{"date-parts":[[2024,6,4]],"date-time":"2024-06-04T05:01:28Z","timestamp":1717477288000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Simultaneously learning intentions and preferences during physical human-robot cooperation"],"prefix":"10.1007","volume":"48","author":[{"given":"Linda","family":"van der Spaa","sequence":"first","affiliation":[]},{"given":"Jens","family":"Kober","sequence":"additional","affiliation":[]},{"given":"Michael","family":"Gienger","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,4]]},"reference":[{"issue":"2","key":"10167_CR1","doi-asserted-by":"publisher","first-page":"61","DOI":"10.3390\/robotics12020061","volume":"12","author":"A Avaei","year":"2023","unstructured":"Avaei, A., van der Spaa, L., Peternel, L., et al. (2023). An incremental inverse reinforcement learning approach for motion planning with human preferences. Robotics, 12(2), 61.","journal-title":"Robotics"},{"key":"10167_CR2","doi-asserted-by":"crossref","unstructured":"Bai, H., Cai, S., & Ye, N., et\u00a0al. (2015). Intention-aware online POMDP planning for autonomous driving in a crowd. In 2015 IEEE international conference on robotics and automation (ICRA) (pp. 454\u2013460). IEEE.","DOI":"10.1109\/ICRA.2015.7139219"},{"key":"10167_CR3","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1016\/B978-0-12-398532-3.00007-5","volume":"7","author":"CL Baker","year":"2014","unstructured":"Baker, C. L., & Tenenbaum, J. B. (2014). Modeling human plan recognition using Bayesian theory of mind. Plan, Activity, and Intent Recognition: Theory and Practice, 7, 177\u2013204.","journal-title":"Plan, Activity, and Intent Recognition: Theory and Practice"},{"key":"10167_CR4","doi-asserted-by":"crossref","unstructured":"Belardinelli, A., Kondapally, A. R., & Ruiken, D., et\u00a0al. (2022). Intention estimation from gaze and motion features for human-robot shared-control object manipulation. In 2022 IEEE\/RSJ international conference on intelligent robots and systems (IROS) (pp. 9806\u20139813). IEEE.","DOI":"10.1109\/IROS47612.2022.9982249"},{"issue":"5","key":"10167_CR5","doi-asserted-by":"publisher","first-page":"497","DOI":"10.1177\/02783649221078031","volume":"41","author":"A Bobu","year":"2022","unstructured":"Bobu, A., Wiggert, M., Tomlin, C., et al. (2022). Inducing structure in reward learning by learning features. The International Journal of Robotics Research, 41(5), 497\u2013518.","journal-title":"The International Journal of Robotics Research"},{"key":"10167_CR6","unstructured":"Boularias, A., Kober, J., & Peters, J. (2011). Relative entropy inverse reinforcement learning. In Proceedings of the 14th international conference on artificial intelligence and statistics (pp. 182\u2013189). IEEE."},{"key":"10167_CR7","doi-asserted-by":"crossref","unstructured":"Buehler, M. C., & Weisswange, T. H. (2018). Online inference of human belief for cooperative robots. In 2018 IEEE\/RSJ international conference on intelligent robots and systems (IROS) (pp. 409\u2013415). IEEE.","DOI":"10.1109\/IROS.2018.8594076"},{"key":"10167_CR8","first-page":"183","volume-title":"Multi-agent reinforcement learning: An overview","author":"L Bu\u015foniu","year":"2010","unstructured":"Bu\u015foniu, L., Babu\u0161ka, R., & De Schutter, B. (2010). Multi-agent reinforcement learning: An overview (pp. 183\u2013221). Springer."},{"key":"10167_CR9","doi-asserted-by":"crossref","unstructured":"Choudhury, R., Swamy, G., & Hadfield-Menell, D., et\u00a0al. (2019). On the utility of model learning in hri. In 2019 14th ACM\/IEEE international conference on human-robot interaction (HRI) (pp. 317\u2013325). IEEE.","DOI":"10.1109\/HRI.2019.8673256"},{"key":"10167_CR10","doi-asserted-by":"crossref","unstructured":"Duchaine, V., & Gosselin, C. M. (2007). General model of human-robot cooperation using a novel velocity based variable impedance control. In Second joint EuroHaptics conference and symposium on haptic interfaces for virtual environment and teleoperator systems (WHC\u201907) (pp. 446\u2013451). IEEE.","DOI":"10.1109\/WHC.2007.59"},{"key":"10167_CR11","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s12984-020-0642-5","volume":"17","author":"NT Fitter","year":"2020","unstructured":"Fitter, N. T., Mohan, M., Kuchenbecker, K. J., et al. (2020). Exercising with Baxter: Preliminary support for assistive social-physical human-robot interaction. Journal of Neuroengineering and Rehabilitation, 17, 1\u201322.","journal-title":"Journal of Neuroengineering and Rehabilitation"},{"key":"10167_CR12","doi-asserted-by":"crossref","unstructured":"Franceschi, P., Maccarini, M., & Piga, D., et\u00a0al. (2023). Human preferences\u2019 optimization in phri collaborative tasks. In 2023 20th international conference on ubiquitous robots (UR) (pp. 693\u2013699). IEEE.","DOI":"10.1109\/UR57808.2023.10202313"},{"key":"10167_CR13","doi-asserted-by":"crossref","unstructured":"Gienger, M., Ruiken, D., & Bates, T., et\u00a0al. (2018). Human-robot cooperative object manipulation with contact changes. In 2018 IEEE\/RSJ international conference on intelligent robots and systems (IROS). IEEE.","DOI":"10.1109\/IROS.2018.8594140"},{"key":"10167_CR14","unstructured":"Hadfield-Menell, D., Russell, S. J., Abbeel, P., et\u00a0al. (2016). Cooperative inverse reinforcement learning. In Advances in neural information processing systems (pp. 3909\u20133917)."},{"key":"10167_CR15","doi-asserted-by":"crossref","unstructured":"Hafs, A., Verdel, D., & Burdet, E., et\u00a0al. (2024). A finite-horizon inverse differential game approach for optimal trajectory-tracking assistance with a wrist exoskeleton. hal-04443499.","DOI":"10.1109\/BioRob60516.2024.10719810"},{"key":"10167_CR16","doi-asserted-by":"crossref","unstructured":"Haninger, K., Hegeler, C., & Peternel, L. (2022). Model predictive control with gaussian processes for flexible multi-modal physical human robot interaction. In 2022 international conference on robotics and automation (ICRA) (pp. 6948\u20136955). IEEE.","DOI":"10.1109\/ICRA46639.2022.9811590"},{"issue":"102","key":"10167_CR17","first-page":"386","volume":"78","author":"A Hanna","year":"2022","unstructured":"Hanna, A., Larsson, S., G\u00f6tvall, P. L., et al. (2022). Deliberative safety for industrial intelligent human\u2013robot collaboration: Regulatory challenges and solutions for taking the next step towards industry 4.0. Robotics and Computer-Integrated Manufacturing, 78(102), 386.","journal-title":"Robotics and Computer-Integrated Manufacturing"},{"key":"10167_CR18","doi-asserted-by":"crossref","unstructured":"Hawkins, K. P., Bansal, S., & Vo, N. N., et\u00a0al. (2014). Anticipating human actions for collaboration in the presence of task and sensor uncertainty. In 2014 IEEE international conference on robotics and automation (ICRA) (pp. 2215\u20132222). IEEE.","DOI":"10.1109\/ICRA.2014.6907165"},{"key":"10167_CR19","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1016\/j.cobeha.2019.04.010","volume":"29","author":"J Jara-Ettinger","year":"2019","unstructured":"Jara-Ettinger, J. (2019). Theory of mind as inverse reinforcement learning. Current Opinion in Behavioral Sciences, 29, 105\u2013110.","journal-title":"Current Opinion in Behavioral Sciences"},{"issue":"1","key":"10167_CR20","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1631\/jzus.C1010010","volume":"12","author":"Jin Zj","year":"2011","unstructured":"Zj, Jin, Qian, H., Sy, Chen, et al. (2011). Convergence analysis of an incremental approach to online inverse reinforcement learning. Journal of Zhejiang University Science C, 12(1), 17\u201324.","journal-title":"Journal of Zhejiang University Science C"},{"key":"10167_CR21","doi-asserted-by":"crossref","unstructured":"Karami, A. B., Jeanpierre, L., & Mouaddib, A. I. (2009). Partially observable markov decision process for managing robot collaboration with human. In 2009 21st IEEE international conference on tools with artificial intelligence (pp. 518\u2013521). IEEE.","DOI":"10.1109\/ICTAI.2009.61"},{"issue":"4","key":"10167_CR22","doi-asserted-by":"publisher","first-page":"3719","DOI":"10.1109\/LRA.2019.2928760","volume":"4","author":"D Koert","year":"2019","unstructured":"Koert, D., Pajarinen, J., Schotschneider, A., et al. (2019). Learning intention aware online adaptation of movement primitives. IEEE Robotics and Automation Letters, 4(4), 3719\u20133726.","journal-title":"IEEE Robotics and Automation Letters"},{"key":"10167_CR23","doi-asserted-by":"publisher","first-page":"453","DOI":"10.1007\/978-3-319-23778-7_30","volume-title":"Experimental robotics","author":"HS Koppula","year":"2016","unstructured":"Koppula, H. S., Jain, A., & Saxena, A. (2016). Anticipatory planning for human-robot teams. Experimental robotics (pp. 453\u2013470). Springer."},{"issue":"2","key":"10167_CR24","doi-asserted-by":"publisher","first-page":"421","DOI":"10.1007\/s10514-021-10030-9","volume":"46","author":"Y Lai","year":"2022","unstructured":"Lai, Y., Paul, G., Cui, Y., et al. (2022). User intent estimation during robot learning using physical human robot interaction primitives. Autonomous Robots, 46(2), 421\u2013436.","journal-title":"Autonomous Robots"},{"issue":"1","key":"10167_CR25","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1177\/02783649211050958","volume":"41","author":"DP Losey","year":"2022","unstructured":"Losey, D. P., Bajcsy, A., O\u2019Malley, M. K., et al. (2022). Physical interaction as communication: Learning robot objectives online from human corrections. The International Journal of Robotics Research, 41(1), 20\u201344.","journal-title":"The International Journal of Robotics Research"},{"key":"10167_CR26","unstructured":"Malik, D., Palaniappan, M., & Fisac, J. F., et\u00a0al. (2018). An efficient, generalized bellman update for cooperative inverse reinforcement learning. arXiv preprintarXiv:1806.03820"},{"key":"10167_CR27","volume-title":"IEEE transactions on robotics","author":"N Mehr","year":"2023","unstructured":"Mehr, N., Wang, M., Bhatt, M., et al. (2023). Maximum-entropy multi-agent dynamic games: Forward and inverse solutions. IEEE transactions on robotics. IEEE."},{"issue":"2","key":"10167_CR28","doi-asserted-by":"publisher","first-page":"10,216","DOI":"10.1016\/j.ifacol.2020.12.2751","volume":"53","author":"S Musi\u0107","year":"2020","unstructured":"Musi\u0107, S., & Hirche, S. (2020). Haptic shared control for human-robot collaboration: A game-theoretical approach. IFAC-PapersOnLine, 53(2), 10,216-10,222.","journal-title":"IFAC-PapersOnLine"},{"issue":"5\u20137","key":"10167_CR29","doi-asserted-by":"publisher","first-page":"618","DOI":"10.1177\/0278364917690593","volume":"36","author":"S Nikolaidis","year":"2017","unstructured":"Nikolaidis, S., Hsu, D., & Srinivasa, S. (2017). Human-robot mutual adaptation in collaborative tasks: Models and experiments. The International Journal of Robotics Research, 36(5\u20137), 618\u2013634.","journal-title":"The International Journal of Robotics Research"},{"key":"10167_CR30","doi-asserted-by":"crossref","unstructured":"Nikolaidis, S., Nath, S., Procaccia, A. D., et\u00a0al. (2017b). Game-theoretic modeling of human adaptation in human-robot collaboration. In Proceedings of the 2017 ACM\/IEEE international conference on human-robot interaction (pp. 323\u2013331).","DOI":"10.1145\/2909824.3020253"},{"key":"10167_CR31","volume-title":"Robotics: Science and systems","author":"SC Ong","year":"2009","unstructured":"Ong, S. C., Png, S. W., Hsu, D., et al. (2009). Pomdps for robotic tasks with mixed observability. Robotics: Science and systems. MIT Press."},{"key":"10167_CR32","doi-asserted-by":"crossref","unstructured":"Parekh, S., Habibian, S., & Losey, D. P. (2022). Rili: Robustly influencing latent intent. In 2022 IEEE\/RSJ international conference on intelligent robots and systems (IROS) (pp. 01\u201308). IEEE.","DOI":"10.1109\/IROS47612.2022.9981601"},{"issue":"1","key":"10167_CR33","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1177\/0278364918812981","volume":"38","author":"JS Park","year":"2019","unstructured":"Park, J. S., Park, C., & Manocha, D. (2019). I-planner: Intention-aware motion planning using learning-based human motion prediction. The International Journal of Robotics Research, 38(1), 23\u201339.","journal-title":"The International Journal of Robotics Research"},{"issue":"10","key":"10167_CR34","doi-asserted-by":"publisher","first-page":"917","DOI":"10.1177\/02783649231182453","volume":"42","author":"L Peters","year":"2023","unstructured":"Peters, L., Rubies-Royo, V., Tomlin, C. J., et al. (2023). Online and offline learning of player objectives from partial observations in dynamic games. The International Journal of Robotics Research, 42(10), 917\u2013937.","journal-title":"The International Journal of Robotics Research"},{"key":"10167_CR35","doi-asserted-by":"crossref","unstructured":"Ranatunga, I., Cremer, S., & Popa, D. O., et\u00a0al. (2015). Intent aware adaptive admittance control for physical human-robot interaction. In 2015 IEEE international conference on robotics and automation (ICRA) (pp. 5635\u20135640). IEEE.","DOI":"10.1109\/ICRA.2015.7139988"},{"issue":"2","key":"10167_CR36","doi-asserted-by":"publisher","first-page":"304","DOI":"10.1109\/TPAMI.2018.2873794","volume":"42","author":"N Rhinehart","year":"2018","unstructured":"Rhinehart, N., & Kitani, K. (2018). First-person activity forecasting from video with online inverse reinforcement learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 42(2), 304\u2013317.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"10167_CR37","unstructured":"Sadigh, D., Sastry, S., & Seshia, S. A., et\u00a0al. (2016). Planning for autonomous cars that leverage effects on human actions. In Robotics: Science and systems, Ann Arbor, MI, USA."},{"key":"10167_CR38","doi-asserted-by":"crossref","unstructured":"Schmerling, E., Leung, K., Vollprecht, W., et\u00a0al. (2018). Multimodal probabilistic model-based planning for human-robot interaction. In 2018 IEEE international conference on robotics and automation (ICRA) (pp. 1\u20139). IEEE.","DOI":"10.1109\/ICRA.2018.8460766"},{"issue":"50","key":"10167_CR39","doi-asserted-by":"publisher","first-page":"24,972","DOI":"10.1073\/pnas.1820676116","volume":"116","author":"W Schwarting","year":"2019","unstructured":"Schwarting, W., Pierson, A., Alonso-Mora, J., et al. (2019). Social behavior for autonomous vehicles. Proceedings of the National Academy of Sciences, 116(50), 24,972-24,978.","journal-title":"Proceedings of the National Academy of Sciences"},{"key":"10167_CR40","doi-asserted-by":"crossref","unstructured":"Schweitzer, P. J., & Seidmann, A. (1985). Generalized polynomial approximations in Markovian decision processes. Journal of Mathematical Analysis and Applications, 110(2), 568\u2013582.","DOI":"10.1016\/0022-247X(85)90317-8"},{"key":"10167_CR41","doi-asserted-by":"crossref","unstructured":"Sendhoff, B., Wersing, H. (2020). Cooperative intelligence-a humane perspective. In 2020 IEEE international conference on human-machine systems (ICHMS) (pp. 1\u20136). IEEE.","DOI":"10.1109\/ICHMS49158.2020.9209387"},{"key":"10167_CR42","doi-asserted-by":"crossref","unstructured":"Shih, A., Ermon, S., & Sadigh, D. (2022). Conditional imitation learning for multi-agent games. In 2022 17th ACM\/IEEE international conference on human-robot interaction (HRI) (pp. 166\u2013175). IEEE.","DOI":"10.1109\/HRI53351.2022.9889671"},{"key":"10167_CR43","unstructured":"Van der Spaa, L., Franzese, G., Kober, J., et\u00a0al. (2022). Disagreement-aware variable impedance control for online learning of physical human-robot cooperation tasks. In ICRA 2022 full day workshop\u2013shared autonomy in physical human-robot interaction: Adaptability and trust."},{"issue":"5","key":"10167_CR44","doi-asserted-by":"publisher","first-page":"1452","DOI":"10.1109\/TRO.2020.2992987","volume":"36","author":"T Stouraitis","year":"2020","unstructured":"Stouraitis, T., Chatzinikolaidis, I., Gienger, M., et al. (2020). Online hybrid motion planning for dyadic collaborative manipulation via bilevel optimization. IEEE Transactions on Robotics, 36(5), 1452\u20131471.","journal-title":"IEEE Transactions on Robotics"},{"key":"10167_CR45","doi-asserted-by":"crossref","unstructured":"Tian, R., Tomizuka, M., Dragan, A. D., et\u00a0al. (2023). Towards modeling and influencing the dynamics of human learning. In Proceedings of the 2023 ACM\/IEEE international conference on human-robot interaction (pp. 350\u2013358).","DOI":"10.1145\/3568162.3578629"},{"key":"10167_CR46","doi-asserted-by":"crossref","unstructured":"Tijsma, A. D., Drugan, M. M., & Wiering, M. A. (2016). Comparing exploration strategies for q-learning in random stochastic mazes. In 2016 IEEE symposium series on computational intelligence (SSCI) (pp. 1\u20138). IEEE.","DOI":"10.1109\/SSCI.2016.7849366"},{"key":"10167_CR47","unstructured":"Wang, W. Z., Shih, A., & Xie, A., et\u00a0al. (2022). Influencing towards stable multi-agent interactions. In Conference on robot learning, PMLR (pp. 1132\u20131143)."},{"key":"10167_CR48","volume-title":"Gaussian processes for machine learning","author":"CK Williams","year":"2006","unstructured":"Williams, C. K., & Rasmussen, C. E. (2006). Gaussian processes for machine learning (Vol. 2). MIT Press."},{"key":"10167_CR49","unstructured":"Xie, A., Losey, D., & Tolsma, R., et\u00a0al. (2021). Learning latent representations to influence multi-agent interaction. In Conference on robot learning (pp. 575\u2013588). PMLR."},{"key":"10167_CR50","doi-asserted-by":"crossref","unstructured":"Zhifei, S., & Joo, E. M. (2012). A review of inverse reinforcement learning theory and recent advances. In Evolutionary computation (CEC), 2012 IEEE congress on, IEEE (pp. 1\u20138).","DOI":"10.1109\/CEC.2012.6256507"},{"key":"10167_CR51","unstructured":"Ziebart, B. D., Maas, A., L. & Bagnell, J. A., et al. (2008). Maximum entropy inverse reinforcement learning. In Aaai (pp. 1433\u20131438). Chicago: IL, USA."},{"issue":"4","key":"10167_CR52","doi-asserted-by":"publisher","first-page":"1584","DOI":"10.1109\/TITS.2019.2913166","volume":"21","author":"A Zyner","year":"2019","unstructured":"Zyner, A., Worrall, S., & Nebot, E. (2019). Naturalistic driver intention and path prediction using recurrent neural networks. IEEE Transactions on Intelligent Transportation Systems., 21(4), 1584\u20131594.","journal-title":"IEEE Transactions on Intelligent Transportation Systems."}],"container-title":["Autonomous Robots"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10514-024-10167-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10514-024-10167-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10514-024-10167-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T01:40:37Z","timestamp":1732153237000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10514-024-10167-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,4]]},"references-count":52,"journal-issue":{"issue":"4-5","published-print":{"date-parts":[[2024,7]]}},"alternative-id":["10167"],"URL":"https:\/\/doi.org\/10.1007\/s10514-024-10167-3","relation":{},"ISSN":["0929-5593","1573-7527"],"issn-type":[{"value":"0929-5593","type":"print"},{"value":"1573-7527","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,6,4]]},"assertion":[{"value":"22 May 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 May 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 June 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}},{"value":"The experimental protocols were approved by the Human Research Ethics Committee at the Delft University of Technology on the 19th of November 2021.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}],"article-number":"11"}}