{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T22:56:54Z","timestamp":1774479414307,"version":"3.50.1"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T00:00:00Z","timestamp":1765324800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T00:00:00Z","timestamp":1765324800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/100012774","name":"Innovation Fund Denmark","doi-asserted-by":"crossref","id":[{"id":"10.13039\/100012774","id-type":"DOI","asserted-by":"crossref"}]},{"name":"ERA PerMed","award":["2021-324"],"award-info":[{"award-number":["2021-324"]}]},{"name":"Austrian Science Fund","award":["10.55776\/I5902"],"award-info":[{"award-number":["10.55776\/I5902"]}]},{"DOI":"10.13039\/501100002347","name":"Bundesministerium f\u00fcr Bildung und Forschung","doi-asserted-by":"publisher","award":["FKZ 01EO1501"],"award-info":[{"award-number":["FKZ 01EO1501"]}],"id":[{"id":"10.13039\/501100002347","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Mitteldeutsche Gesellschaft f\u00fcr Pneumologie","award":["2018-MDGP-PA-002"],"award-info":[{"award-number":["2018-MDGP-PA-002"]}]},{"name":"European Commission Research Directorate-General","award":["779282"],"award-info":[{"award-number":["779282"]}]},{"name":"Progetto PETictCAC, Regione Toscana"},{"name":"Saxon State Ministry for Science, Culture and Tourism"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["AI Ethics"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s43681-025-00837-y","type":"journal-article","created":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T10:45:48Z","timestamp":1765363548000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Hyper-selective explainability: an empirical case study of the utility of explainability in a clinical decision support system"],"prefix":"10.1007","volume":"6","author":[{"given":"Shaul A.","family":"Duke","sequence":"first","affiliation":[]},{"given":"Peter","family":"Sand\u00f8e","sequence":"additional","affiliation":[]},{"given":"Thomas B\u00f8ker","family":"Lund","sequence":"additional","affiliation":[]},{"given":"Elisabetta Maria","family":"Abenavoli","sequence":"additional","affiliation":[]},{"given":"Thomas","family":"Beyer","sequence":"additional","affiliation":[]},{"given":"Daria","family":"Ferrara","sequence":"additional","affiliation":[]},{"given":"Armin","family":"Frille","sequence":"additional","affiliation":[]},{"given":"Stefan","family":"Gruenert","sequence":"additional","affiliation":[]},{"given":"Osama","family":"Sabri","sequence":"additional","affiliation":[]},{"given":"Roberto","family":"Sciagr\u00e0","sequence":"additional","affiliation":[]},{"given":"Miriam","family":"Pepponi","sequence":"additional","affiliation":[]},{"given":"Hesse","family":"Swen","sequence":"additional","affiliation":[]},{"given":"Anke","family":"T\u00f6njes","sequence":"additional","affiliation":[]},{"given":"Hubert","family":"Wirtz","sequence":"additional","affiliation":[]},{"given":"Josef","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Lalith Kumar","family":"Shiyam Sundar","sequence":"additional","affiliation":[]},{"given":"Sune","family":"Holm","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,10]]},"reference":[{"key":"837_CR1","doi-asserted-by":"crossref","first-page":"329","DOI":"10.1136\/medethics-2021-107531","volume":"47","author":"JM Dur\u00e1n","year":"2021","unstructured":"Dur\u00e1n, J.M., Jongsma, K.R.: Who is afraid of black box algorithms? On the epistemological and ethical basis of trust in medical AI. J. Med. Ethics 47, 329\u2013335 (2021)","journal-title":"J. Med. Ethics"},{"key":"837_CR2","doi-asserted-by":"crossref","first-page":"6","DOI":"10.1136\/jme-2022-108814","volume":"50","author":"F Funer","year":"2024","unstructured":"Funer, F., et al.: Responsibility and decision-making authority in using clinical decision support systems: an empirical-ethical exploration of German prospective professionals\u2019 preferences and concerns. J. Med. Ethics 50, 6\u201311 (2024)","journal-title":"J. Med. Ethics"},{"key":"837_CR3","first-page":"1","volume":"12","author":"E Hildt","year":"2025","unstructured":"Hildt, E.: What is the role of explainability in medical artificial intelligence? a case-based approach. Bioeng 12, 1\u201325 (2025)","journal-title":"Bioeng"},{"key":"837_CR4","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s12911-020-01332-6","volume":"20","author":"J Amann","year":"2020","unstructured":"Amann, J., Blasimme, A., Vayena, E., Frey, D., Madai, V.I.: Explainability for artificial intelligence in healthcare: a multidisciplinary perspective. BMC Med. Inform. Decis. Mak. 20, 1\u20139 (2020)","journal-title":"BMC Med. Inform. Decis. Mak."},{"key":"837_CR5","doi-asserted-by":"crossref","first-page":"284","DOI":"10.1126\/science.abg1834","volume":"373","author":"B Babic","year":"2021","unstructured":"Babic, B., Gerke, S., Evgeniou, T., Cohen, I.G.: Beware explanations from AI in health care. Science 373, 284\u2013286 (2021)","journal-title":"Science"},{"key":"837_CR6","doi-asserted-by":"crossref","first-page":"15","DOI":"10.1002\/hast.973","volume":"49","author":"AJ London","year":"2019","unstructured":"London, A.J.: Artificial intelligence and black-box medical decisions: accuracy versus explainability. Hastings Cent. Rep. 49, 15\u201321 (2019)","journal-title":"Hastings Cent. Rep."},{"key":"837_CR7","doi-asserted-by":"crossref","first-page":"1","DOI":"10.2196\/25187","volume":"23","author":"S Muralitharan","year":"2021","unstructured":"Muralitharan, S., et al.: Machine learning\u2013based early warning systems for clinical deterioration: systematic scoping review. J. Med. Internet Res. 23, 1\u201322 (2021)","journal-title":"J. Med. Internet Res."},{"key":"837_CR8","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1371\/journal.pdig.0000016","volume":"1","author":"J Amann","year":"2022","unstructured":"Amann, J., et al.: To explain or not to explain?: artificial intelligence explainability in clinical decision support systems. PLOS digit. health 1, 1\u201318 (2022)","journal-title":"PLOS digit. health"},{"key":"837_CR9","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s12910-024-01103-2","volume":"25","author":"N Freyer","year":"2024","unstructured":"Freyer, N., Gro\u00df, D., Lipprandt, M.: The ethical requirement of explainability for AI-DSS in healthcare: a systematic review of reasons. BMC Med. Ethics 25, 1\u201311 (2024)","journal-title":"BMC Med. Ethics"},{"key":"837_CR10","unstructured":"Tonekaboni, S., Joshi, S., McCradden, M.M., Goldenberg, A.: What clinicians want: contextualizing explainable machine learning for clinical end use. Paper presented at: machine learning for healthcare conference. (2019) August 9 2019, Ann Arbor, MI"},{"key":"837_CR11","doi-asserted-by":"crossref","first-page":"231","DOI":"10.1136\/bmjqs-2018-008370","volume":"28","author":"R Challen","year":"2019","unstructured":"Challen, R., et al.: Artificial intelligence, bias and clinical safety. BMJ Qual. Saf. 28, 231\u2013237 (2019)","journal-title":"BMJ Qual. Saf."},{"issue":"3","key":"837_CR12","doi-asserted-by":"publisher","first-page":"1155","DOI":"10.1287\/mnsc.2016.2643","volume":"64","author":"BJ Dietvorst","year":"2018","unstructured":"Dietvorst, B.J., Simmons, J.P., Massey, C.: Overcoming algorithm aversion: people will use imperfect algorithms if they can (even slightly) modify them. Manage. Sci. 64(3), 1155\u20131170 (2018). https:\/\/doi.org\/10.1287\/mnsc.2016.2643","journal-title":"Manage. Sci."},{"key":"837_CR13","doi-asserted-by":"crossref","first-page":"209","DOI":"10.1111\/jopp.12262","volume":"30","author":"K Vredenburgh","year":"2022","unstructured":"Vredenburgh, K.: The right to explanation. J Polit Philos 30, 209\u2013229 (2022)","journal-title":"J Polit Philos"},{"key":"837_CR14","doi-asserted-by":"crossref","first-page":"825","DOI":"10.1111\/puar.13293","volume":"81","author":"M Busuioc","year":"2021","unstructured":"Busuioc, M.: Accountable artificial intelligence: holding algorithms to account. Public Adm. Rev. 81, 825\u2013836 (2021)","journal-title":"Public Adm. Rev."},{"key":"837_CR15","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3236009","volume":"51","author":"R Guidotti","year":"2019","unstructured":"Guidotti, R., et al.: A survey of methods for explaining black box models. ACM Comput. Surv. 51, 1\u201342 (2019)","journal-title":"ACM Comput. Surv."},{"key":"837_CR16","first-page":"1","volume":"2022","author":"PN Srinivasu","year":"2022","unstructured":"Srinivasu, P.N., Sandhya, N., Jhaveri, R.H., Raut, R.: From blackbox to explainable AI in healthcare: existing tools and case studies. Mob. Inf. Syst. 2022, 1\u201320 (2022)","journal-title":"Mob. Inf. Syst."},{"key":"837_CR17","first-page":"1","volume":"15","author":"JR Zech","year":"2018","unstructured":"Zech, J.R., et al.: Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: a cross-sectional study. PLoS Med. 15, 1\u201317 (2018)","journal-title":"PLoS Med."},{"key":"837_CR18","doi-asserted-by":"crossref","first-page":"745","DOI":"10.1016\/S2589-7500(21)00208-9","volume":"3","author":"M Ghassemi","year":"2021","unstructured":"Ghassemi, M., Oakden-Rayner, L., Beam, A.L.: The false hope of current approaches to explainable artificial intelligence in health care. Lancet Digit. Health. 3, 745\u2013750 (2021)","journal-title":"Lancet Digit. Health."},{"key":"837_CR19","first-page":"633","volume":"165","author":"JA Kroll","year":"2017","unstructured":"Kroll, J.A., et al.: Accountable algorithms. U. Pa. L. Rev. 165, 633\u2013705 (2017)","journal-title":"U. Pa. L. Rev."},{"key":"837_CR20","doi-asserted-by":"crossref","first-page":"206","DOI":"10.1038\/s42256-019-0048-x","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1, 206\u2013215 (2019)","journal-title":"Nat. Mach. Intell."},{"key":"837_CR21","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1038\/s41746-020-0254-2","volume":"3","author":"CM Cutillo","year":"2020","unstructured":"Cutillo, C.M., et al.: Machine intelligence in healthcare: perspectives on trustworthiness, explainability, usability, and transparency. npj Digit Med. 3, 1\u20135 (2020)","journal-title":"npj Digit Med."},{"key":"837_CR22","doi-asserted-by":"crossref","first-page":"437","DOI":"10.1136\/medethics-2020-106922","volume":"47","author":"A Ferrario","year":"2021","unstructured":"Ferrario, A., Loi, M., Vigan\u00f2, E.: Trust does not need to be human: it is possible to trust medical AI. J. Med. Ethics 47, 437\u2013438 (2021)","journal-title":"J. Med. Ethics"},{"key":"837_CR23","doi-asserted-by":"publisher","unstructured":"Holm, S.H.: On the justified use of AI decision support in evidence-based medicine: validity, explainability, and responsibility. Camb Q Healthc Ethics. 1\u20137 (2023). https:\/\/doi.org\/10.1017\/S0963180123000294","DOI":"10.1017\/S0963180123000294"},{"key":"837_CR24","first-page":"169","volume-title":"Handbook of artificial intelligence in healthcare: Vol 2: practicalities and prospects volume 2","author":"J Gerlings","year":"2022","unstructured":"Gerlings, J., Jensen, M.S., Shollo, A.: Explainable AI, but explainable to whom? An exploratory case study of xAI in healthcare. In: Lim, C., Chen, Y., Vaidya, A., Mahorkar, C., Jain, L.C. (eds.) Handbook of artificial intelligence in healthcare: Vol 2: practicalities and prospects volume 2, pp. 169\u2013198. Springer-Verlag, London (2022)"},{"key":"837_CR25","doi-asserted-by":"crossref","first-page":"341","DOI":"10.1136\/medethics-2021-107352","volume":"47","author":"A Mishra","year":"2021","unstructured":"Mishra, A.: Transparent AI: reliabilist and proud. J. Med. Ethics 47, 341\u2013342 (2021)","journal-title":"J. Med. Ethics"},{"key":"837_CR26","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.artint.2022.103839","volume":"316","author":"O Wysocki","year":"2023","unstructured":"Wysocki, O., et al.: Assessing the communication gap between AI models and healthcare professionals: explainability, utility and trust in AI-driven clinical decision-making. Artif. Intell. 316, 1\u201317 (2023)","journal-title":"Artif. Intell."},{"key":"837_CR27","doi-asserted-by":"crossref","DOI":"10.1016\/j.artmed.2022.102423","volume":"133","author":"C Combi","year":"2022","unstructured":"Combi, C., et al.: A manifesto on explainability for artificial intelligence in medicine. Artif. Intell. Med. 133, 102423 (2022)","journal-title":"Artif. Intell. Med."},{"key":"837_CR28","volume-title":"Fairness and machine learning: limitations and opportunities","author":"S Barocas","year":"2019","unstructured":"Barocas, S., Hardt, M., Narayanan, A.: Fairness and machine learning: limitations and opportunities. ABC-CLIO, Santa Barbara, CA (2019)"},{"key":"837_CR29","doi-asserted-by":"crossref","first-page":"525","DOI":"10.1007\/s13347-017-0293-z","volume":"31","author":"PB de Laat","year":"2018","unstructured":"de Laat, P.B.: Algorithmic decision-making based on machine learning from big data: can transparency restore accountability? Philos. Technol. 31, 525\u2013541 (2018)","journal-title":"Philos. Technol."},{"key":"837_CR30","doi-asserted-by":"crossref","first-page":"449","DOI":"10.1007\/s43681-021-00091-y","volume":"2","author":"T Izumo","year":"2022","unstructured":"Izumo, T., Weng, Y.-H.: Coarse ethics: how to ethically assess explainable artificial intelligence. AI Ethics 2, 449\u2013461 (2022)","journal-title":"AI Ethics"},{"key":"837_CR31","doi-asserted-by":"crossref","first-page":"815","DOI":"10.1007\/s43681-022-00142-y","volume":"2","author":"D Vale","year":"2022","unstructured":"Vale, D., El-Sharif, A., Ali, M.: Explainable artificial intelligence (XAI) post-hoc explainability methods: risks and limitations in non-discrimination law. AI Ethics 2, 815\u2013826 (2022)","journal-title":"AI Ethics"},{"key":"837_CR32","doi-asserted-by":"crossref","first-page":"2735","DOI":"10.1007\/s43681-024-00586-4","volume":"5","author":"J Fritz","year":"2025","unstructured":"Fritz, J.: On the scope of the right to explanation. AI Ethics. 5, 2735\u20132747 (2025)","journal-title":"AI Ethics."},{"key":"837_CR33","doi-asserted-by":"crossref","DOI":"10.1201\/9781003530244","volume-title":"Debiasing AI: rethinking the intersection of innovation and sustainability","author":"DH Shin","year":"2025","unstructured":"Shin, D.H.: Debiasing AI: rethinking the intersection of innovation and sustainability. Routledge, London (2025)"},{"key":"837_CR34","doi-asserted-by":"crossref","first-page":"478","DOI":"10.1136\/medethics-2019-105935","volume":"46","author":"JJ Hatherley","year":"2020","unstructured":"Hatherley, J.J.: Limits of trust in medical AI. J. Med. Ethics 46, 478\u2013481 (2020)","journal-title":"J. Med. Ethics"},{"key":"837_CR35","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1007\/s43681-022-00152-w","volume":"3","author":"JF Cortese","year":"2023","unstructured":"Cortese, J.F., Cozman, F.G., Lucca-Silveira, M.P., Bechara, A.F.: Should explainability be a fifth ethical principle in AI ethics? AI Ethics 3, 1\u201312 (2023)","journal-title":"AI Ethics"},{"key":"837_CR36","first-page":"251","volume":"24","author":"C Kemper","year":"2019","unstructured":"Kemper, C.: Kafkaesque AI? Legal decision-making in the era of machine learning. Intell. Prop. L. J. 24, 251\u2013294 (2019)","journal-title":"Intell. Prop. L. J."},{"key":"837_CR37","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1001\/jamanetworkopen.2020.32320","volume":"4","author":"SN Blomberg","year":"2021","unstructured":"Blomberg, S.N., et al.: Effect of machine learning on dispatcher recognition of out-of-hospital cardiac arrest during calls to emergency medical services: a randomized clinical trial. JAMA Netw. Open 4, 1\u201310 (2021)","journal-title":"JAMA Netw. Open"},{"key":"837_CR38","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10676-022-09627-0","volume":"24","author":"SA Duke","year":"2022","unstructured":"Duke, S.A.: Deny, dismiss and downplay: developers\u2019 attitudes towards risk and their role in risk creation in the field of healthcare-AI. Ethics Inf. Technol. 24, 1 (2022). https:\/\/doi.org\/10.1007\/s10676-022-09627-0","journal-title":"Ethics Inf. Technol."},{"key":"837_CR39","doi-asserted-by":"crossref","first-page":"257","DOI":"10.1007\/s43681-022-00161-9","volume":"3","author":"A Fernandez-Quilez","year":"2023","unstructured":"Fernandez-Quilez, A.: Deep learning in radiology: ethics of data and on the value of algorithm transparency, interpretability and explainability. AI Ethics. 3, 257\u2013265 (2023)","journal-title":"AI Ethics."},{"key":"837_CR40","doi-asserted-by":"crossref","first-page":"78","DOI":"10.1109\/MC.2021.3092610","volume":"54","author":"A Holzinger","year":"2021","unstructured":"Holzinger, A., M\u00fcller, H.: Toward human\u2013AI interfaces to support explainability and causability in medical AI. Computer 54, 78\u201386 (2021)","journal-title":"Computer"},{"key":"837_CR41","doi-asserted-by":"crossref","first-page":"3015","DOI":"10.1007\/s43681-024-00622-3","volume":"5","author":"S Alpsancar","year":"2025","unstructured":"Alpsancar, S., Buhl, H.M., Matzner, T., Scharlau, I.: Explanation needs and ethical demands: unpacking the instrumental value of XAI. AI Ethics. 5, 3015\u20133033 (2025)","journal-title":"AI Ethics."},{"key":"837_CR42","doi-asserted-by":"crossref","first-page":"219","DOI":"10.1007\/s43681-021-00121-9","volume":"2","author":"C Herzog","year":"2022","unstructured":"Herzog, C.: On the risk of confusing interpretability with explicability. AI Ethics. 2, 219\u2013225 (2022)","journal-title":"AI Ethics."},{"key":"837_CR43","first-page":"1","volume":"9","author":"C Borch","year":"2022","unstructured":"Borch, C., Hee Min, B.: Toward a sociology of machine learning explainability: human\u2013machine interaction in deep neural network-based automated trading. BD&amp;S 9, 1\u201313 (2022)","journal-title":"BD&amp;S"},{"key":"837_CR44","doi-asserted-by":"crossref","first-page":"205","DOI":"10.1136\/medethics-2019-105586","volume":"46","author":"T Grote","year":"2020","unstructured":"Grote, T., Berens, P.: On the ethics of algorithmic decision-making in healthcare. J. Med. Ethics 46, 205 (2020)","journal-title":"J. Med. Ethics"},{"key":"837_CR45","doi-asserted-by":"publisher","DOI":"10.1002\/jcsm.13571","author":"D Ferrara","year":"2024","unstructured":"Ferrara, D., et al.: Detection of cancer-associated cachexia in lung cancer patients using whole-body [18f]FDG-PET\/CT imaging: a multi-centre study. J. Cachexia. Sarcopenia Muscle (2024). https:\/\/doi.org\/10.1002\/jcsm.13571","journal-title":"J. Cachexia. Sarcopenia Muscle"},{"key":"837_CR46","doi-asserted-by":"publisher","DOI":"10.1007\/s00259-024-06689-8","author":"A Frille","year":"2024","unstructured":"Frille, A., et al.: \u201cMetabolic fingerprints\u201d of cachexia in lung cancer patients. Eur. J. Nucl. Med. Mol. Imaging (2024). https:\/\/doi.org\/10.1007\/s00259-024-06689-8","journal-title":"Eur. J. Nucl. Med. Mol. Imaging"},{"key":"837_CR47","doi-asserted-by":"crossref","first-page":"2898","DOI":"10.1016\/j.clnu.2021.02.005","volume":"40","author":"M Muscaritoli","year":"2021","unstructured":"Muscaritoli, M., et al.: ESPEN practical guideline: clinical nutrition in cancer. Clin. Nutr. 40, 2898\u20132913 (2021)","journal-title":"Clin. Nutr."},{"key":"837_CR48","doi-asserted-by":"publisher","first-page":"5597","DOI":"10.2147\/CMAR.S261585","volume":"12","author":"J Ni","year":"2020","unstructured":"Ni, J., Zhang, L.: Cancer cachexia: definition, staging, and emerging treatments. Cancer Manag. Res. 12, 5597\u20135605 (2020). https:\/\/doi.org\/10.2147\/CMAR.S261585","journal-title":"Cancer Manag. Res."},{"key":"837_CR49","first-page":"64","volume":"4","author":"V Suhag","year":"2015","unstructured":"Suhag, V., Sunita, B.S., Sarin, A., Singh, A.K.: Cancer, malnutrition and cachexia: we must break the triad. Int. J. Med. Phys. Clin. Eng. Radiat. 4, 64\u201370 (2015)","journal-title":"Int. J. Med. Phys. Clin. Eng. Radiat."},{"key":"837_CR50","doi-asserted-by":"crossref","first-page":"489","DOI":"10.1016\/S1470-2045(10)70218-7","volume":"12","author":"K Fearon","year":"2011","unstructured":"Fearon, K., et al.: Definition and classification of cancer cachexia: an international consensus. Lancet Oncol. 12, 489\u2013495 (2011)","journal-title":"Lancet Oncol."},{"key":"837_CR51","doi-asserted-by":"crossref","first-page":"80","DOI":"10.1177\/160940690600500107","volume":"5","author":"J Fereday","year":"2006","unstructured":"Fereday, J., Muir-Cochrane, E.: Demonstrating rigor using thematic analysis: a hybrid approach of inductive and deductive coding and theme development. Int J Qual Methods 5, 80\u201392 (2006)","journal-title":"Int J Qual Methods"},{"key":"837_CR52","doi-asserted-by":"crossref","first-page":"783","DOI":"10.1177\/1049732319889354","volume":"30","author":"N Bergen","year":"2020","unstructured":"Bergen, N., Labont\u00e9, R.: \u201cEverything is perfect, and we have no problems\u201d: detecting and limiting social desirability bias in qualitative research. Qual. Health Res. 30, 783\u2013792 (2020)","journal-title":"Qual. Health Res."}],"container-title":["AI and Ethics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-025-00837-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s43681-025-00837-y","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-025-00837-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T10:27:25Z","timestamp":1774261645000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s43681-025-00837-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,10]]},"references-count":52,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["837"],"URL":"https:\/\/doi.org\/10.1007\/s43681-025-00837-y","relation":{},"ISSN":["2730-5953","2730-5961"],"issn-type":[{"value":"2730-5953","type":"print"},{"value":"2730-5961","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,10]]},"assertion":[{"value":"20 September 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 November 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interests"}},{"value":"The University of Copenhagen research ethics review board approved our interviews (approval: 504\u20130358\/22\u20135000) on October 10, 2022. Respondents gave recorded consent before starting interviews.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval and consent to participate"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}],"article-number":"53"}}