{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T19:22:47Z","timestamp":1770146567095,"version":"3.49.0"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2023,6,3]],"date-time":"2023-06-03T00:00:00Z","timestamp":1685750400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,6,3]],"date-time":"2023-06-03T00:00:00Z","timestamp":1685750400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100000995","name":"Australian National University","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100000995","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["AI &amp; Soc"],"published-print":{"date-parts":[[2024,10]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Artificially intelligent systems will be used to make increasingly important decisions about us. Many of these decisions will have to be made without universal agreement about the relevant moral facts. For other kinds of disagreement, it is at least usually obvious what kind of solution is called for. What makes moral disagreement especially challenging is that there are three different ways of handling it. <jats:italic>Moral solutions<\/jats:italic> apply a moral theory or related principles and largely ignore the details of the disagreement. <jats:italic>Compromise solutions<\/jats:italic> apply a method of finding a compromise and taking information about the disagreement as input. <jats:italic>Epistemic solutions<\/jats:italic> apply an evidential rule that treats the details of the disagreement as evidence of moral truth. Proposals for all three kinds of solutions can be found in the AI ethics and value alignment literature, but little has been said to justify choosing one over the other. I argue that the choice is best framed in terms of <jats:italic>moral risk<\/jats:italic>.\n<\/jats:p>","DOI":"10.1007\/s00146-023-01697-y","type":"journal-article","created":{"date-parts":[[2023,6,3]],"date-time":"2023-06-03T19:01:41Z","timestamp":1685818901000},"page":"2425-2438","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Moral disagreement and artificial intelligence"],"prefix":"10.1007","volume":"39","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1663-5970","authenticated-orcid":false,"given":"Pamela","family":"Robinson","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,3]]},"reference":[{"issue":"3","key":"1697_CR1","doi-asserted-by":"publisher","first-page":"530","DOI":"10.2307\/1600416","volume":"66","author":"L Alexander","year":"1999","unstructured":"Alexander L (1999) \u201cWith Me, It\u2019s All er Nuthin\u2019\u201d: formalism in law and morality. Univ Chicago Law Rev 66(3):530\u2013565. https:\/\/doi.org\/10.2307\/1600416","journal-title":"Univ Chicago Law Rev"},{"key":"1697_CR2","unstructured":"Anderson M, Andersen SL, Armen C (2006) MedEthEx: a prototype medical ethics advisor. In: Proceedings of the 18th conference on innovative applications of artificial intelligence, vol 2, pp 1759\u20131765. AAAI Press, Boston"},{"key":"1697_CR3","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1038\/s41586-018-0637-6","volume":"563","author":"E Awad","year":"2018","unstructured":"Awad E, Dsouza S, Kim R, Schulz J, Henrich J, Shariff A, Bonnefon F, Rahwan I (2018) The moral machine experiment. Nature 563:59\u201364. https:\/\/doi.org\/10.1038\/s41586-018-0637-6","journal-title":"Nature"},{"issue":"1","key":"1697_CR4","doi-asserted-by":"publisher","first-page":"165","DOI":"10.1007\/s00146-017-0760-1","volume":"35","author":"S Baum","year":"2020","unstructured":"Baum S (2020) Social choice ethics in artificial intelligence. AI Soc 35(1):165\u2013176. https:\/\/doi.org\/10.1007\/s00146-017-0760-1","journal-title":"AI Soc"},{"key":"1697_CR5","doi-asserted-by":"publisher","first-page":"167","DOI":"10.5040\/9781472594150.ch-009","volume-title":"Advances in experimental moral psychology","author":"JR Beebe","year":"2014","unstructured":"Beebe JR (2014) How different kinds of disagreement impact folk metaethical judgments. In: Sarkissian H, Wright JC (eds) Advances in experimental moral psychology. Bloomsbury Academic, London, pp 167\u2013187"},{"key":"1697_CR6","doi-asserted-by":"publisher","DOI":"10.1093\/oso\/9780190652951.003.001","volume-title":"Robot ethics 2.0: from autonomous cars to artificial intelligence","author":"V Bhargava","year":"2017","unstructured":"Bhargava V, Kim TW (2017) Autonomous vehicles and moral uncertainty. In: Lin P, Abney K, Jenkins R (eds) Robot ethics 2.0: from autonomous cars to artificial intelligence. Oxford University Press, New York. https:\/\/doi.org\/10.1093\/oso\/9780190652951.003.001"},{"issue":"4","key":"1697_CR7","doi-asserted-by":"publisher","first-page":"591","DOI":"10.1007\/s11023-017-9448-z","volume":"27","author":"K Bogosian","year":"2017","unstructured":"Bogosian K (2017) Implementation of moral uncertainty in intelligent machines. Mind Mach 27(4):591\u2013608. https:\/\/doi.org\/10.1007\/s11023-017-9448-z","journal-title":"Mind Mach"},{"key":"1697_CR9","volume-title":"Superintelligence: paths, dangers, strategies","author":"N Bostrom","year":"2014","unstructured":"Bostrom N (2014) Superintelligence: paths, dangers, strategies. Oxford University Press, Oxford"},{"key":"1697_CR10","doi-asserted-by":"publisher","first-page":"316","DOI":"10.1017\/CBO9781139046855.020","volume-title":"The cambridge handbook of artificial intelligence","author":"N Bostrom","year":"2014","unstructured":"Bostrom N, Yudkowsky E (2014) The ethics of artificial intelligence. In: Frankish K, Ramsey WM (eds) The cambridge handbook of artificial intelligence. Cambridge University Press, Cambridge, pp 316\u2013334. https:\/\/doi.org\/10.1017\/CBO9781139046855.020"},{"key":"1697_CR11","doi-asserted-by":"publisher","unstructured":"Brennan-Marquez K, Chiao V (2021) Algorithmic decision-making when humans disagree on ends. New Crim Law Rev 24(3):275\u2013300. https:\/\/doi.org\/10.1525\/nclr.2021.24.3.275","DOI":"10.1525\/nclr.2021.24.3.275"},{"key":"1697_CR61","doi-asserted-by":"publisher","unstructured":"Brink O (1984) Moral realism and the sceptical arguments from disagreement and queerness. Australasian Journal of Philosophy 62(2):111\u2013125. https:\/\/doi.org\/10.1080\/00048408412341311","DOI":"10.1080\/00048408412341311"},{"issue":"3","key":"1697_CR12","doi-asserted-by":"publisher","first-page":"355","DOI":"10.1080\/0952813X.2014.895108","volume":"26","author":"M Brundage","year":"2014","unstructured":"Brundage M (2014) Limitations and risks of machine ethics. J Exp Theor Artif Intell 26(3):355\u2013372. https:\/\/doi.org\/10.1080\/0952813X.2014.895108","journal-title":"J Exp Theor Artif Intell"},{"key":"1697_CR13","doi-asserted-by":"publisher","DOI":"10.1007\/s11245-018-9607-8","author":"J Carlson","year":"2018","unstructured":"Carlson J (2018) Epistemology of disagreement, bias, and political deliberation: the problems for a conciliary democracy. Topoi. https:\/\/doi.org\/10.1007\/s11245-018-9607-8","journal-title":"Topoi"},{"key":"1697_CR15","unstructured":"Ecoffet A, Lehman J (2021) Reinforcement learning under moral uncertainty. Proceedings of the 38th International Conference on Machine Learning, pp 2926\u20132936"},{"key":"1697_CR16","volume-title":"Routledge handbook of political epistemology","author":"E Edenberg","year":"2021","unstructured":"Edenberg E (2021) Political disagreement: epistemic or civic peers? In: Hannon M, de Ridder J (eds) Routledge handbook of political epistemology. Routledge, London"},{"key":"1697_CR17","doi-asserted-by":"publisher","first-page":"132","DOI":"10.1093\/oso\/9780198801221.003.0007","volume-title":"Oxford studies in political philosophy","author":"D Enoch","year":"2017","unstructured":"Enoch D (2017) Political philosophy and epistemology: the case of public reason. In: Sobel D, Vallentyne P, Wall S (eds) Oxford studies in political philosophy, vol 3. Oxford University Press, Oxford, pp 132\u2013165. https:\/\/doi.org\/10.1093\/oso\/9780198801221.003.0007"},{"key":"1697_CR18","doi-asserted-by":"publisher","first-page":"403","DOI":"10.1007\/s10892-017-9252-2","volume":"21","author":"A Etzioni","year":"2017","unstructured":"Etzioni A, Etzioni O (2017) Incorporating ethics into artificial intelligence. J Ethics 21:403\u2013418. https:\/\/doi.org\/10.1007\/s10892-017-9252-2","journal-title":"J Ethics"},{"key":"1697_CR19","doi-asserted-by":"publisher","DOI":"10.1007\/s00146-020-01089-6","author":"P Formosa","year":"2020","unstructured":"Formosa P, Ryan M (2020) Making moral machines: why we need artificial moral agents. AI Soc. https:\/\/doi.org\/10.1007\/s00146-020-01089-6","journal-title":"AI Soc"},{"key":"1697_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2020.103261","author":"R Freedman","year":"2020","unstructured":"Freedman R, Schaich Borg J, Sinnott-Armstrong W, Dickerson JP, Conitzer V (2020) Adapting a kidney exchange algorithm to align with human values. Artif Intell. https:\/\/doi.org\/10.1016\/j.artint.2020.103261","journal-title":"Artif Intell"},{"key":"1697_CR21","doi-asserted-by":"publisher","first-page":"411","DOI":"10.1007\/s11023-020-09539-2","volume":"30","author":"I Gabriel","year":"2020","unstructured":"Gabriel I (2020) Artificial intelligence, values, and alignment. Mind Mach 30:411\u2013437. https:\/\/doi.org\/10.1007\/s11023-020-09539-2","journal-title":"Mind Mach"},{"issue":"3","key":"1697_CR22","doi-asserted-by":"publisher","first-page":"311","DOI":"10.1093\/arisoc\/aow018","volume":"116","author":"H Greaves","year":"2016","unstructured":"Greaves H (2016) Cluelessness. Proc Aristot Soc 116(3):311\u2013339.\u00a0https:\/\/doi.org\/10.1093\/arisoc\/aow018","journal-title":"Proc Aristot Soc"},{"issue":"1","key":"1697_CR23","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1086\/293260","volume":"101","author":"A Gutmann","year":"1990","unstructured":"Gutmann A, Thompson D (1990) Moral conflict and political consensus. Ethics 101(1):64\u201388. https:\/\/doi.org\/10.1086\/293260","journal-title":"Ethics"},{"key":"1697_CR24","volume-title":"Democracy and disagreement","author":"A Gutmann","year":"1996","unstructured":"Gutmann A, Thompson D (1996) Democracy and disagreement. Cambridge University Press, Cambridge"},{"key":"1697_CR25","doi-asserted-by":"publisher","first-page":"669","DOI":"10.1007\/s10677-018-9896-4","volume":"21","author":"J Himmelreich","year":"2018","unstructured":"Himmelreich J (2018) Never mind the trolley: the ethics of autonomous vehicles in mundane situations. Ethical Theory Moral Pract 21:669\u2013684. https:\/\/doi.org\/10.1007\/s10677-018-9896-4","journal-title":"Ethical Theory Moral Pract"},{"issue":"1","key":"1697_CR26","doi-asserted-by":"publisher","first-page":"75","DOI":"10.1177\/1470594X17729132","volume":"17","author":"K Kappel","year":"2018","unstructured":"Kappel K (2018) How moral disagreement may ground principled moral compromise. Politics Philos Econ 17(1):75\u201396. https:\/\/doi.org\/10.1177\/1470594X17729132","journal-title":"Politics Philos Econ"},{"key":"1697_CR27","doi-asserted-by":"publisher","first-page":"107","DOI":"10.1007\/s10676-017-9419-3","volume":"19","author":"D Leben","year":"2017","unstructured":"Leben D (2017) A rawlsian algorithm for autonomous vehicles. Ethics Inf Technol 19:107\u2013115. https:\/\/doi.org\/10.1007\/s10676-017-9419-3","journal-title":"Ethics Inf Technol"},{"key":"1697_CR28","volume-title":"Oxford handbook of deliberative democracy","author":"C List","year":"2018","unstructured":"List C (2018) Democratic deliberation and social choice: a review. In: B\u00e4chtiger A et al (eds) Oxford handbook of deliberative democracy. Oxford University Press, Oxford"},{"key":"1697_CR29","doi-asserted-by":"publisher","DOI":"10.1093\/oso\/9780195126105.001.0001","volume-title":"Moral uncertainty and its consequences","author":"T Lockhart","year":"2000","unstructured":"Lockhart T (2000) Moral uncertainty and its consequences. Oxford University Press, Oxford"},{"issue":"3","key":"1697_CR30","doi-asserted-by":"publisher","first-page":"508","DOI":"10.1086\/669564","volume":"123","author":"W MacAskill","year":"2013","unstructured":"MacAskill W (2013) The infectiousness of nihilism. Ethics 123(3):508\u2013520.\u00a0https:\/\/doi.org\/10.1086\/669564","journal-title":"Ethics"},{"key":"1697_CR31","unstructured":"MacAskill W (2014) Normative uncertainty. PhD dissertation, Department of Philosophy, Oxford University, Oxford"},{"issue":"500","key":"1697_CR32","doi-asserted-by":"publisher","first-page":"967","DOI":"10.1093\/mind\/fzv169","volume":"125","author":"W MacAskill","year":"2016","unstructured":"MacAskill W (2016) Normative uncertainty as a voting problem. Mind 125(500):967\u20131004. https:\/\/doi.org\/10.1093\/mind\/fzv169","journal-title":"Mind"},{"issue":"2","key":"1697_CR33","doi-asserted-by":"publisher","first-page":"327","DOI":"10.1111\/nous.12264","volume":"54","author":"W MacAskill","year":"2020","unstructured":"MacAskill W, Ord T (2020) Why maximize expected choice-worthiness? No\u00fbs 54(2):327\u2013353.\u00a0https:\/\/doi.org\/10.1111\/nous.12264","journal-title":"No\u00fbs"},{"key":"1697_CR34","doi-asserted-by":"publisher","DOI":"10.1093\/oso\/9780198722274.001.0001","volume-title":"Moral uncertainty","author":"W MacAskill","year":"2020","unstructured":"MacAskill W, Bykvist K, Ord T (2020) Moral uncertainty. Oxford University Press, Oxford"},{"key":"1697_CR35","volume-title":"Ethics: inventing right and wrong","author":"JL Mackie","year":"1977","unstructured":"Mackie JL (1977) Ethics: inventing right and wrong. Penguin Books, Harmondsworth"},{"key":"1697_CR36","doi-asserted-by":"publisher","unstructured":"Martinho A, Kroesen M, Chorus C (2021) An empirical approach to capture moral uncertainty in AI. Minds & Machines 31:215\u2013237. https:\/\/doi.org\/10.1007\/s11023-021-09556-9","DOI":"10.1007\/s11023-021-09556-9"},{"key":"1697_CR37","doi-asserted-by":"publisher","first-page":"87","DOI":"10.1093\/oso\/9780199542062.003.0005","volume-title":"Oxford studies in metaethics","author":"S McGrath","year":"2008","unstructured":"McGrath S (2008) Moral disagreement and moral expertise. In: Shafer-Landau R (ed) Oxford studies in metaethics, vol 3. Oxford, New York, pp 87\u2013108"},{"issue":"7","key":"1697_CR38","doi-asserted-by":"publisher","first-page":"1871","DOI":"10.1007\/s11098-016-0825-x","volume":"174","author":"R Muldoon","year":"2017","unstructured":"Muldoon R (2017) Exploring tradeoffs in accommodating moral diversity. Philos Stud 174(7):1871\u20131883. https:\/\/doi.org\/10.1007\/s11098-016-0825-x","journal-title":"Philos Stud"},{"key":"1697_CR39","volume-title":"What is pluralism?","author":"T Mulligan","year":"2020","unstructured":"Mulligan T (2020) Social choice or collective decision-making: what is politics all about? In: Kaul V, Salvatore I (eds) What is pluralism? Routledge India, London"},{"key":"1697_CR40","doi-asserted-by":"crossref","unstructured":"Noothigattu R, Gaikwad SS, Awad E, Dsouza S, Rahwan I, Ravikumar P, Procaccia A (2018) A voting-based system for ethical decision making. Paper presented at the thirty-second AAAI conference on artificial intelligence, New Orleans, Louisiana, February 2\u20138","DOI":"10.1609\/aaai.v32i1.11512"},{"key":"1697_CR41","doi-asserted-by":"publisher","DOI":"10.1080\/13698230.2020.1737475","author":"I O\u2019Flynn","year":"2020","unstructured":"O\u2019Flynn I, Set\u00e4l\u00e4 M (2020) Deliberative disagreement and compromise. Crit Rev Int Soc Pol Phil. https:\/\/doi.org\/10.1080\/13698230.2020.1737475","journal-title":"Crit Rev Int Soc Pol Phil"},{"key":"1697_CR42","doi-asserted-by":"publisher","first-page":"413","DOI":"10.1093\/oso\/9780190905033.003.0015","volume-title":"Ethics of artificial intelligence","author":"S Petersen","year":"2020","unstructured":"Petersen S (2020) Machines learning values. In: Liao SM (ed) Ethics of artificial intelligence. Oxford, New York, pp 413\u2013435"},{"key":"1697_CR43","doi-asserted-by":"publisher","first-page":"291","DOI":"10.1201\/9781351251389-21","volume-title":"Artificial intelligence safety and security","author":"M Prasad","year":"2018","unstructured":"Prasad M (2018) Social choice and the value alignment problem. In: Yampolsky RV (ed) Artificial intelligence safety and security. Chapman and Hall, London, pp 291\u2013314"},{"key":"1697_CR44","volume-title":"Political liberalism","author":"J Rawls","year":"2005","unstructured":"Rawls J (2005) Political liberalism, Expanded. Columbia University Press, New York","edition":"Expanded"},{"key":"1697_CR45","volume-title":"The right and the good","author":"WD Ross","year":"1930","unstructured":"Ross WD (1930) The right and the good. Oxford University Press, Oxford"},{"key":"1697_CR46","doi-asserted-by":"publisher","first-page":"742","DOI":"10.1086\/505234","volume":"116","author":"J Ross","year":"2006","unstructured":"Ross J (2006) Rejecting ethical deflationism. Ethics 116:742\u2013768.\u00a0https:\/\/doi.org\/10.1086\/505234","journal-title":"Ethics"},{"key":"1697_CR47","unstructured":"Russell S (2019) Human compatible: AI and the problem of control. Penguin"},{"key":"1697_CR48","volume-title":"Oxford studies in metaethics","author":"A Sepielli","year":"2009","unstructured":"Sepielli A (2009) What to do when you don\u2019t know what to do. In: Shafer-Landau R (ed) Oxford studies in metaethics. Oxford University Press, Oxford"},{"key":"1697_CR49","doi-asserted-by":"publisher","unstructured":"Sinnott-Armstrong W, Skorburg JA (2021) How AI can aid bioethics. Journal of Practical Ethics 9(1). https:\/\/doi.org\/10.3998\/jpe.1175","DOI":"10.3998\/jpe.1175"},{"key":"1697_CR50","volume-title":"The epistemology of group disagreement","author":"M Skipper","year":"2020","unstructured":"Skipper M, Steglich-Petersen A (2021) When conciliation frustrates the epistemic priorities of groups. In: Broncano-Berrocal F, Carter JA (eds) The epistemology of group disagreement. Routledge, New York"},{"issue":"1","key":"1697_CR51","doi-asserted-by":"publisher","first-page":"37","DOI":"10.1080\/23294515.2019.1706206","volume":"11","author":"JA Skorburg","year":"2020","unstructured":"Skorburg JA, Sinnott-Armstrong W, Conitzer V (2020) AI methods in bioethics. AJOB Empirical Bioethics 11(1):37\u201339. https:\/\/doi.org\/10.1080\/23294515.2019.1706206","journal-title":"AJOB Empirical Bioethics"},{"issue":"6","key":"1697_CR52","doi-asserted-by":"publisher","DOI":"10.1111\/phc3.12493","volume":"13","author":"F Tersman","year":"2018","unstructured":"Tersman F (2018) Recent work on reflective equilibrium and method in ethics. Philos Compass 13(6):e12493. https:\/\/doi.org\/10.1111\/phc3.12493","journal-title":"Philos Compass"},{"key":"1697_CR53","volume-title":"Principled sentencing and artificial intelligence","author":"F Thomsen","year":"2023","unstructured":"Thomsen F (2022) Iudicium ex machinae: the ethical challenges of automated decision-making in criminal sentencing. In: Roberts J, Ryberg J (eds) Sentencing and artificial intelligence. Oxford University Press, Oxford"},{"issue":"3","key":"1697_CR54","doi-asserted-by":"publisher","first-page":"610","DOI":"10.1086\/292869","volume":"97","author":"W Tolhurst","year":"1987","unstructured":"Tolhurst W (1987) The argument from moral disagreement. Ethics 97(3):610\u2013621. https:\/\/doi.org\/10.1086\/292869","journal-title":"Ethics"},{"key":"1697_CR55","doi-asserted-by":"publisher","first-page":"27","DOI":"10.1007\/s10676-017-9440-6","volume":"20","author":"P Vamplew","year":"2018","unstructured":"Vamplew P, Dazeley R, Foale C, Firmin S, Mummery J (2018) Human-aligned artificial intelligence is a multiobjective problem. Ethics Inf Technol 20:27\u201340. https:\/\/doi.org\/10.1007\/s10676-017-9440-6","journal-title":"Ethics Inf Technol"},{"issue":"4","key":"1697_CR57","doi-asserted-by":"publisher","first-page":"486","DOI":"10.1111\/jopp.12152","volume":"26","author":"H van Wietmarschen","year":"2018","unstructured":"van Wietmarschen H (2018) Reasonable citizens and epistemic peers: a skeptical problem for political liberalism. J Political Philos 26(4):486\u2013507. https:\/\/doi.org\/10.1111\/jopp.12152","journal-title":"J Political Philos"},{"issue":"4","key":"1697_CR58","doi-asserted-by":"publisher","first-page":"763","DOI":"10.1086\/293447","volume":"102","author":"DB Wong","year":"1992","unstructured":"Wong DB (1992) Coping with moral conflict and ambiguity. Ethics 102(4):763\u2013784. https:\/\/doi.org\/10.1086\/293447","journal-title":"Ethics"},{"key":"1697_CR59","doi-asserted-by":"publisher","first-page":"225","DOI":"10.1007\/s13347-019-00355","volume":"33","author":"P-H Wong","year":"2020","unstructured":"Wong P-H (2020) Democratizing algorithmic fairness. Philos Technol 33:225\u2013244. https:\/\/doi.org\/10.1007\/s13347-019-00355","journal-title":"Philos Technol"},{"issue":"1","key":"1697_CR60","doi-asserted-by":"publisher","first-page":"2237","DOI":"10.1609\/aaai.v33i01.33012237","volume":"33","author":"H Zhang","year":"2019","unstructured":"Zhang H, Conitzer V (2019) A PAC framework for aggregating agents\u2019 judgments. Proc AAAI Conf Artif Intell 33(1):2237\u20132244. https:\/\/doi.org\/10.1609\/aaai.v33i01.33012237","journal-title":"Proc AAAI Conf Artif Intell"}],"container-title":["AI &amp; SOCIETY"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00146-023-01697-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00146-023-01697-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00146-023-01697-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,20]],"date-time":"2024-09-20T04:07:27Z","timestamp":1726805247000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00146-023-01697-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,3]]},"references-count":58,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["1697"],"URL":"https:\/\/doi.org\/10.1007\/s00146-023-01697-y","relation":{},"ISSN":["0951-5666","1435-5655"],"issn-type":[{"value":"0951-5666","type":"print"},{"value":"1435-5655","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,6,3]]},"assertion":[{"value":"10 May 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 May 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 June 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"On behalf of all authors, the corresponding author states that there is no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}