{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T02:57:56Z","timestamp":1776135476320,"version":"3.50.1"},"reference-count":61,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2023,9,14]],"date-time":"2023-09-14T00:00:00Z","timestamp":1694649600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,9,14]],"date-time":"2023-09-14T00:00:00Z","timestamp":1694649600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/100016017","name":"Duke-NUS Medical School","doi-asserted-by":"crossref","id":[{"id":"10.13039\/100016017","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Estate of Tan Sri Khoo Teck Puat"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["npj Digit. Med."],"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Artificial intelligence (AI) has demonstrated the ability to extract insights from data, but the fairness of such data-driven insights remains a concern in high-stakes fields. Despite extensive developments, issues of AI fairness in clinical contexts have not been adequately addressed. A fair model is normally expected to perform equally across subgroups defined by sensitive variables (e.g., age, gender\/sex, race\/ethnicity, socio-economic status, etc.). Various fairness measurements have been developed to detect differences between subgroups as evidence of bias, and bias mitigation methods are designed to reduce the differences detected. This perspective of fairness, however, is misaligned with some key considerations in clinical contexts. The set of sensitive variables used in healthcare applications must be carefully examined for relevance and justified by clear clinical motivations. In addition, clinical AI fairness should closely investigate the ethical implications of fairness measurements (e.g., potential conflicts between group- and individual-level fairness) to select suitable and objective metrics. Generally defining AI fairness as \u201cequality\u201d is not necessarily reasonable in clinical settings, as differences may have clinical justifications and do not indicate biases. Instead, \u201cequity\u201d would be an appropriate objective of clinical AI fairness. Moreover, clinical feedback is essential to developing fair and well-performing AI models, and efforts should be made to actively involve clinicians in the process. The adaptation of AI fairness towards healthcare is not self-evident due to misalignments between technical developments and clinical considerations. Multidisciplinary collaboration between AI researchers, clinicians, and ethicists is necessary to bridge the gap and translate AI fairness into real-life benefits.<\/jats:p>","DOI":"10.1038\/s41746-023-00918-4","type":"journal-article","created":{"date-parts":[[2023,9,14]],"date-time":"2023-09-14T18:03:47Z","timestamp":1694714627000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":63,"title":["A translational perspective towards clinical AI fairness"],"prefix":"10.1038","volume":"6","author":[{"given":"Mingxuan","family":"Liu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6758-4472","authenticated-orcid":false,"given":"Yilin","family":"Ning","sequence":"additional","affiliation":[]},{"given":"Salinelat","family":"Teixayavong","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9883-9167","authenticated-orcid":false,"given":"Mayli","family":"Mertens","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Daniel Shu Wei","family":"Ting","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1068-7868","authenticated-orcid":false,"given":"Lionel Tim-Ee","family":"Cheng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6916-5960","authenticated-orcid":false,"given":"Jasmine Chiat Ling","family":"Ong","sequence":"additional","affiliation":[]},{"given":"Zhen Ling","family":"Teo","sequence":"additional","affiliation":[]},{"given":"Ting Fang","family":"Tan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7365-0053","authenticated-orcid":false,"given":"Narrendar","family":"RaviChandran","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9459-9461","authenticated-orcid":false,"given":"Fei","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6712-6626","authenticated-orcid":false,"given":"Leo Anthony","family":"Celi","sequence":"additional","affiliation":[]},{"given":"Marcus Eng Hock","family":"Ong","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3610-4883","authenticated-orcid":false,"given":"Nan","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,14]]},"reference":[{"key":"918_CR1","doi-asserted-by":"crossref","first-page":"433","DOI":"10.1093\/mind\/LIX.236.433","volume":"59","author":"AM Turing","year":"1950","unstructured":"Turing, A. M. Computing machinery and intelligence. Mind 59, 433\u2013460 (1950).","journal-title":"Mind"},{"key":"918_CR2","doi-asserted-by":"crossref","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y. & Hinton, G. Deep learning. Nature 521, 436\u2013444 (2015).","journal-title":"Nature"},{"key":"918_CR3","doi-asserted-by":"crossref","first-page":"5","DOI":"10.1177\/0008125619864925","volume":"61","author":"M Haenlein","year":"2019","unstructured":"Haenlein, M. & Kaplan, A. A brief history of artificial intelligence: On the past, present, and future of artificial intelligence. Calif. Manag. Rev. 61, 5\u201314 (2019).","journal-title":"Calif. Manag. Rev."},{"key":"918_CR4","unstructured":"OpenAI. ChatGPT (Mar 14 version) [Large language model], https:\/\/chat.openai.com\/chat (2023)."},{"key":"918_CR5","doi-asserted-by":"crossref","first-page":"1201","DOI":"10.1056\/NEJMra2302038","volume":"388","author":"CJ Haug","year":"2023","unstructured":"Haug, C. J. & Drazen, J. M. Artificial Intelligence and Machine Learning in Clinical Medicine, 2023. N. Engl. J. Med. 388, 1201\u20131208 (2023).","journal-title":"N. Engl. J. Med."},{"key":"918_CR6","doi-asserted-by":"crossref","unstructured":"Bohr, A. & Memarzadeh, K. The rise of artificial intelligence in healthcare applications. Artificial Intelligence in Healthcare, 25\u201360 (2020).","DOI":"10.1016\/B978-0-12-818438-7.00002-2"},{"key":"918_CR7","unstructured":"Mertens, M. Bias in Medicine. The Rowman & Littlefield Handbook of Bioethics. 103\u2013117 (Rowman & Littlefield, 2022)."},{"key":"918_CR8","doi-asserted-by":"crossref","first-page":"561802","DOI":"10.3389\/frai.2020.561802","volume":"3","author":"RR Fletcher","year":"2020","unstructured":"Fletcher, R. R., Nakeshimana, A. & Olubeko, O. Addressing Fairness, Bias, and Appropriate Use of Artificial Intelligence and Machine Learning in Global Health. Front Artif. Intell. 3, 561802 (2020).","journal-title":"Front Artif. Intell."},{"key":"918_CR9","volume":"5","author":"TC Tsai","year":"2022","unstructured":"Tsai, T. C. et al. Algorithmic fairness in pandemic forecasting: lessons from COVID-19. npj Digital Med. 5, 59 (2022).","journal-title":"npj Digital Med."},{"key":"918_CR10","doi-asserted-by":"crossref","first-page":"874","DOI":"10.1056\/NEJMms2004740","volume":"383","author":"DA Vyas","year":"2020","unstructured":"Vyas, D. A., Eisenstein, L. G. & Jones, D. S. Hidden in plain sight \u2014 reconsidering the use of race correction in clinical algorithms. N. Engl. J. Med. 383, 874\u2013882 (2020).","journal-title":"N. Engl. J. Med."},{"key":"918_CR11","doi-asserted-by":"crossref","first-page":"e1343","DOI":"10.1097\/TXD.0000000000001343","volume":"8","author":"MD Doshi","year":"2022","unstructured":"Doshi, M. D., Schaubel, D. E., Xu, Y., Rao, P. S. & Sung, R. S. Clinical utility in adopting race-free kidney donor risk index. Transpl. Direct 8, e1343 (2022).","journal-title":"Transpl. Direct"},{"key":"918_CR12","doi-asserted-by":"crossref","first-page":"1996","DOI":"10.1038\/s41591-022-01961-6","volume":"28","author":"V Volovici","year":"2022","unstructured":"Volovici, V., Syn, N. L., Ercole, A., Zhao, J. J. & Liu, N. Steps to avoid overuse and misuse of machine learning in clinical research. Nat. Med. 28, 1996\u20131999 (2022).","journal-title":"Nat. Med."},{"key":"918_CR13","doi-asserted-by":"crossref","DOI":"10.1038\/s41746-020-0288-5","volume":"3","author":"D Cirillo","year":"2020","unstructured":"Cirillo, D. et al. Sex and gender differences and biases in artificial intelligence for biomedicine and healthcare. npj Digital Med. 3, 81 (2020).","journal-title":"npj Digital Med."},{"key":"918_CR14","doi-asserted-by":"crossref","first-page":"1515","DOI":"10.1111\/ajt.16986","volume":"22","author":"JC Lai","year":"2022","unstructured":"Lai, J. C., Pomfret, E. A. & Verna, E. C. Implicit bias and the gender inequity in liver transplantation. Am. J. Transpl. 22, 1515\u20131518 (2022).","journal-title":"Am. J. Transpl."},{"key":"918_CR15","doi-asserted-by":"publisher","unstructured":"Menezes, H. F., Ferreira, A. S. C., Pereira, E. T. & Gomes, H. M. Bias and Fairness in Face Detection. 2021 34th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), 247\u2013254 (2021). https:\/\/doi.org\/10.1109\/SIBGRAPI54419.2021.00041.","DOI":"10.1109\/SIBGRAPI54419.2021.00041"},{"key":"918_CR16","unstructured":"Angwin, J., Larson, J., Mattu, S. & Kirchner, L. Machine bias: There\u2019s software used across the country to predict future criminals. and it\u2019s biased against blacks, https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing (2016)."},{"key":"918_CR17","doi-asserted-by":"publisher","unstructured":"Caton, S. & Haas, C. Fairness in Machine Learning: A Survey. ACM Comput. Surv. (2023). https:\/\/doi.org\/10.1145\/3616865.","DOI":"10.1145\/3616865"},{"key":"918_CR18","doi-asserted-by":"publisher","unstructured":"Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K. & Galstyan, A. A Survey on Bias and Fairness in Machine Learning. ACM Comput. Surv. 54 (2021). https:\/\/doi.org\/10.1145\/3457607.","DOI":"10.1145\/3457607"},{"key":"918_CR19","doi-asserted-by":"crossref","first-page":"141","DOI":"10.1146\/annurev-statistics-042720-125902","volume":"8","author":"S Mitchell","year":"2021","unstructured":"Mitchell, S., Potash, E., Barocas, S., D\u2019Amour, A. & Lum, K. Algorithmic fairness: choices, assumptions, and definitions. Annu. Rev. Stat. Appl. 8, 141\u2013163 (2021).","journal-title":"Annu. Rev. Stat. Appl."},{"key":"918_CR20","doi-asserted-by":"crossref","first-page":"2020","DOI":"10.1093\/jamia\/ocaa094","volume":"27","author":"M DeCamp","year":"2020","unstructured":"DeCamp, M. & Lindvall, C. Latent bias and the implementation of artificial intelligence in medicine. J. Am. Med Inf. Assoc. 27, 2020\u20132023 (2020).","journal-title":"J. Am. Med Inf. Assoc."},{"key":"918_CR21","doi-asserted-by":"crossref","first-page":"e213909","DOI":"10.1001\/jamanetworkopen.2021.3909","volume":"4","author":"Y Park","year":"2021","unstructured":"Park, Y. et al. Comparison of methods to reduce bias from clinical prediction models of postpartum depression. JAMA Netw. Open 4, e213909\u2013e213909 (2021).","journal-title":"JAMA Netw. Open"},{"key":"918_CR22","doi-asserted-by":"crossref","first-page":"619","DOI":"10.1176\/ps.62.6.pss6206_0619","volume":"62","author":"KB Kozhimannil","year":"2011","unstructured":"Kozhimannil, K. B., Trinacty, C. M., Busch, A. B., Huskamp, H. A. & Adams, A. S. Racial and ethnic disparities in postpartum depression care among low-income women. Psychiatr. Serv. 62, 619\u2013625 (2011).","journal-title":"Psychiatr. Serv."},{"key":"918_CR23","volume":"5","author":"AAH de Hond","year":"2022","unstructured":"de Hond, A. A. H. et al. Guidelines and quality criteria for artificial intelligence-based prediction models in healthcare: a scoping review. npj Digital Med. 5, 2 (2022).","journal-title":"npj Digital Med."},{"key":"918_CR24","doi-asserted-by":"crossref","first-page":"841","DOI":"10.1126\/science.1193032","volume":"329","author":"G Genovese","year":"2010","unstructured":"Genovese, G. et al. Association of trypanolytic ApoL1 variants with kidney disease in African Americans. Science 329, 841\u2013845 (2010).","journal-title":"Science"},{"key":"918_CR25","doi-asserted-by":"crossref","first-page":"896","DOI":"10.1038\/s42256-021-00396-x","volume":"3","author":"KT Rodolfa","year":"2021","unstructured":"Rodolfa, K. T., Lamba, H. & Ghani, R. Empirical observation of negligible fairness\u2013accuracy trade-offs in machine learning for public policy. Nat. Mach. Intell. 3, 896\u2013904 (2021).","journal-title":"Nat. Mach. Intell."},{"key":"918_CR26","doi-asserted-by":"publisher","unstructured":"Xu, J. et al. Algorithmic fairness in computational medicine. eBioMedicine 84 (2022). https:\/\/doi.org\/10.1016\/j.ebiom.2022.104250.","DOI":"10.1016\/j.ebiom.2022.104250"},{"key":"918_CR27","doi-asserted-by":"publisher","unstructured":"Dwork, C., Hardt, M., Pitassi, T., Reingold, O. & Zemel, R. Fairness through Awareness. In Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, 214\u2013226 (2012). https:\/\/doi.org\/10.1145\/2090236.2090255.","DOI":"10.1145\/2090236.2090255"},{"key":"918_CR28","doi-asserted-by":"publisher","unstructured":"Kusner, M., Loftus, J., Russell, C. & Silva, R. Counterfactual fairness. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) 30, 4067\u20134077 (2017). https:\/\/doi.org\/10.24963\/ijcai.2019\/199.","DOI":"10.24963\/ijcai.2019\/199"},{"key":"918_CR29","doi-asserted-by":"crossref","first-page":"302","DOI":"10.1109\/TIT.2019.2945779","volume":"66","author":"D Russo","year":"2020","unstructured":"Russo, D. & Zou, J. How Much Does Your Data Exploration Overfit? Controlling Bias via Information Usage. IEEE Trans. Inf. Theory 66, 302\u2013323 (2020).","journal-title":"IEEE Trans. Inf. Theory"},{"key":"918_CR30","doi-asserted-by":"publisher","unstructured":"Puyol-Ant\u00f3n, E. et al. Fairness in Cardiac MR Image Analysis: An Investigation of Bias Due to Data Imbalance in Deep Learning Based Segmentation. Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021: 24th International Conference, Strasbourg, France, September 27\u2013October 1, 2021, Proceedings, Part III, 413\u2013423 (2021). https:\/\/doi.org\/10.1007\/978-3-030-87199-4_39.","DOI":"10.1007\/978-3-030-87199-4_39"},{"key":"918_CR31","doi-asserted-by":"crossref","DOI":"10.1186\/1743-0003-6-31","volume":"6","author":"AA Butler","year":"2009","unstructured":"Butler, A. A., Menant, J. C., Tiedemann, A. C. & Lord, S. R. Age and gender differences in seven tests of functional mobility. J. Neuroeng. Rehabilitation 6, 31 (2009).","journal-title":"J. Neuroeng. Rehabilitation"},{"key":"918_CR32","doi-asserted-by":"crossref","first-page":"529","DOI":"10.1007\/s43681-021-00067-y","volume":"1","author":"MSA Lee","year":"2021","unstructured":"Lee, M. S. A., Floridi, L. & Singh, J. Formalising trade-offs beyond algorithmic fairness: lessons from ethical philosophy and welfare economics. AI Ethics 1, 529\u2013544 (2021).","journal-title":"AI Ethics"},{"key":"918_CR33","doi-asserted-by":"publisher","unstructured":"Binns, R. On the Apparent Conflict between Individual and Group Fairness. In Proceedings of the 2020 ACM Conference on Fairness, Accountability, and Transparency (FAccT \u201820), 514\u2013524 (2020). https:\/\/doi.org\/10.1145\/3351095.3372864.","DOI":"10.1145\/3351095.3372864"},{"key":"918_CR34","doi-asserted-by":"crossref","first-page":"104525","DOI":"10.1016\/j.ebiom.2023.104525","volume":"90","author":"AB Mbakwe","year":"2023","unstructured":"Mbakwe, A. B., Lourentzou, I., Celi, L. A. & Wu, J. T. Fairness metrics for health AI: we have a long way to go. EBioMedicine 90, 104525 (2023).","journal-title":"EBioMedicine"},{"key":"918_CR35","doi-asserted-by":"publisher","unstructured":"Kleinberg, J., Mullainathan, S. & Raghavan, M. Inherent Trade-Offs in the Fair Determination of Risk Scores. In Proceedings of 8th Innovations in Theoretical Computer Science Conference (ITCS 2017), 43:41\u201343:23 (2017). https:\/\/doi.org\/10.4230\/LIPICS.ITCS.2017.43.","DOI":"10.4230\/LIPICS.ITCS.2017.43"},{"key":"918_CR36","volume":"6","author":"J Yang","year":"2023","unstructured":"Yang, J., Soltan, A. A. S., Eyre, D. W., Yang, Y. & Clifton, D. A. An adversarial training framework for mitigating algorithmic biases in clinical machine learning. npj Digital Med. 6, 55 (2023).","journal-title":"npj Digital Med."},{"key":"918_CR37","doi-asserted-by":"publisher","unstructured":"Hardt, M., Price, E., Price, E. & Srebro, N. Equality of Opportunity in Supervised Learning. In Proceedings of the 30th International Conference on Neural Information Processing Systems 29 (2016). https:\/\/doi.org\/10.5555\/3157382.3157469.","DOI":"10.5555\/3157382.3157469"},{"key":"918_CR38","doi-asserted-by":"publisher","unstructured":"DiCiccio, C., Vasudevan, S., Basu, K., Kenthapadi, K. & Agarwal, D. Evaluating Fairness Using Permutation Tests. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (Kdd \u201820), 1467-1477 (2020). https:\/\/doi.org\/10.1145\/3394486.3403199.","DOI":"10.1145\/3394486.3403199"},{"key":"918_CR39","doi-asserted-by":"publisher","unstructured":"Taskesen, B., Blanchet, J., Kuhn, D. & Nguyen, V. A. A Statistical Test for Probabilistic Fairness. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (FAccT \u201821), 648-665 (2021). https:\/\/doi.org\/10.1145\/3442188.3445927.","DOI":"10.1145\/3442188.3445927"},{"key":"918_CR40","doi-asserted-by":"crossref","first-page":"332","DOI":"10.1037\/0022-006X.67.3.332","volume":"67","author":"AE Kazdin","year":"1999","unstructured":"Kazdin, A. E. The meanings and measurement of clinical significance. J. Consult Clin. Psychol. 67, 332\u2013339 (1999).","journal-title":"J. Consult Clin. Psychol."},{"key":"918_CR41","doi-asserted-by":"publisher","unstructured":"Biswas, A. & Mukherjee, S. Ensuring Fairness under Prior Probability Shifts. In Proceedings of the 2021 AAAI\/ACM Conference on AI, Ethics, and Society, 414\u2013424 (2021). https:\/\/doi.org\/10.1145\/3461702.3462596.","DOI":"10.1145\/3461702.3462596"},{"key":"918_CR42","doi-asserted-by":"crossref","first-page":"104","DOI":"10.1177\/0002716206294796","volume":"609","author":"D Pager","year":"2007","unstructured":"Pager, D. The Use of Field Experiments for Studies of Employment Discrimination: Contributions, Critiques, and Directions for the Future. Ann. Am. Acad. Political Soc. Sci. 609, 104\u2013133 (2007).","journal-title":"Ann. Am. Acad. Political Soc. Sci."},{"key":"918_CR43","doi-asserted-by":"crossref","first-page":"35","DOI":"10.1080\/15265161.2023.2186516","volume":"23","author":"A-FJ de Kanter","year":"2023","unstructured":"de Kanter, A.-F. J., van Daal, M., de Graeff, N. & Jongsma, K. R. Preventing Bias in Medical Devices: Identifying Morally Significant Differences. Am. J. Bioeth. 23, 35\u201337 (2023).","journal-title":"Am. J. Bioeth."},{"key":"918_CR44","doi-asserted-by":"crossref","first-page":"210100","DOI":"10.1183\/16000617.0100-2021","volume":"31","author":"M Ragavan","year":"2022","unstructured":"Ragavan, M. & Patel, M. I. The evolving landscape of sex-based differences in lung cancer: a distinct disease in women. Eur. Resp. Rev. 31, 210100 (2022).","journal-title":"Eur. Resp. Rev."},{"key":"918_CR45","first-page":"1397","volume":"97","author":"A Cotter","year":"2019","unstructured":"Cotter, A. et al. Training well-generalizing classifiers for fairness metrics and other data-dependent constraints. Proc. 36th Int. Conf. Mach. Learn. 97, 1397\u20131405 (2019).","journal-title":"Proc. 36th Int. Conf. Mach. Learn."},{"key":"918_CR46","volume":"20","author":"CWQ Ng","year":"2020","unstructured":"Ng, C. W. Q., Lim, J. N. W., Liu, J. & Hartman, M. Presentation of breast cancer, help seeking behaviour and experience of patients in their cancer journey in Singapore: a qualitative study. BMC Cancer 20, 1080 (2020).","journal-title":"BMC Cancer"},{"key":"918_CR47","first-page":"Article 115","volume":"54","author":"N Mehrabi","year":"2021","unstructured":"Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K. & Galstyan, A. A Survey on Bias and Fairness in Machine Learning. ACM Comput. Surv. 54, Article 115 (2021).","journal-title":"ACM Comput. Surv."},{"key":"918_CR48","doi-asserted-by":"crossref","first-page":"447","DOI":"10.1126\/science.aax2342","volume":"366","author":"Z Obermeyer","year":"2019","unstructured":"Obermeyer, Z., Powers, B., Vogeli, C. & Mullainathan, S. Dissecting racial bias in an algorithm used to manage the health of populations. Science 366, 447\u2013453 (2019).","journal-title":"Science"},{"key":"918_CR49","doi-asserted-by":"crossref","first-page":"194","DOI":"10.1097\/TP.0000000000000969","volume":"100","author":"BI Freedman","year":"2016","unstructured":"Freedman, B. I. et al. APOL1 genotype and kidney transplantation outcomes from deceased African American Donors. Transplantation 100, 194\u2013202 (2016).","journal-title":"Transplantation"},{"key":"918_CR50","doi-asserted-by":"crossref","first-page":"878","DOI":"10.1016\/j.chest.2022.05.006","volume":"162","author":"JH Brems","year":"2022","unstructured":"Brems, J. H., Ferryman, K., McCormack, M. C. & Sugarman, J. Ethical considerations regarding the use of race in pulmonary function testing. CHEST 162, 878\u2013881 (2022).","journal-title":"CHEST"},{"key":"918_CR51","doi-asserted-by":"crossref","first-page":"2176","DOI":"10.1038\/s41591-021-01595-0","volume":"27","author":"L Seyyed-Kalantari","year":"2021","unstructured":"Seyyed-Kalantari, L., Zhang, H., McDermott, M. B. A., Chen, I. Y. & Ghassemi, M. Underdiagnosis bias of artificial intelligence algorithms applied to chest radiographs in under-served patient populations. Nat. Med. 27, 2176\u20132182 (2021).","journal-title":"Nat. Med."},{"key":"918_CR52","volume":"11","author":"Q Zhao","year":"2020","unstructured":"Zhao, Q., Adeli, E. & Pohl, K. M. Training confounder-free deep learning models for medical applications. Nat. Commun. 11, 6010 (2020).","journal-title":"Nat. Commun."},{"key":"918_CR53","doi-asserted-by":"crossref","first-page":"922","DOI":"10.1136\/medethics-2020-106636","volume":"48","author":"M Mertens","year":"2022","unstructured":"Mertens, M., King, O. C., Putten, M. J. A. M. V. & Boenink, M. Can we learn from hidden mistakes? Self-fulfilling prophecy and responsible neuroprognostic innovation. J. Med. Ethics 48, 922\u2013928 (2022).","journal-title":"J. Med. Ethics"},{"key":"918_CR54","doi-asserted-by":"crossref","first-page":"173","DOI":"10.1093\/phe\/phu042","volume":"8","author":"MJ Smith","year":"2015","unstructured":"Smith, M. J. Health equity in public health: clarifying our commitment. Public Health Ethics 8, 173\u2013184 (2015).","journal-title":"Public Health Ethics"},{"key":"918_CR55","doi-asserted-by":"crossref","first-page":"254","DOI":"10.1136\/jech.57.4.254","volume":"57","author":"P Braveman","year":"2003","unstructured":"Braveman, P. & Gruskin, S. Defining equity in health. J. Epidemiol. Community Health 57, 254 (2003).","journal-title":"J. Epidemiol. Community Health"},{"key":"918_CR56","doi-asserted-by":"crossref","first-page":"491","DOI":"10.1093\/jamia\/ocz192","volume":"27","author":"S Reddy","year":"2020","unstructured":"Reddy, S., Allan, S., Coghlan, S. & Cooper, P. A governance model for the application of AI in health care. J. Am. Med Inf. Assoc. 27, 491\u2013497 (2020).","journal-title":"J. Am. Med Inf. Assoc."},{"key":"918_CR57","doi-asserted-by":"crossref","first-page":"1631","DOI":"10.1093\/jamia\/ocac078","volume":"29","author":"AD Bedoya","year":"2022","unstructured":"Bedoya, A. D. et al. A framework for the oversight and local deployment of safe and high-quality prediction models. J. Am. Med Inf. Assoc. 29, 1631\u20131636 (2022).","journal-title":"J. Am. Med Inf. Assoc."},{"key":"918_CR58","doi-asserted-by":"crossref","first-page":"e048008","DOI":"10.1136\/bmjopen-2020-048008","volume":"11","author":"GS Collins","year":"2021","unstructured":"Collins, G. S. et al. Protocol for development of a reporting guideline (TRIPOD-AI) and risk of bias tool (PROBAST-AI) for diagnostic and prognostic prediction model studies based on artificial intelligence. BMJ Open 11, e048008 (2021).","journal-title":"BMJ Open"},{"key":"918_CR59","doi-asserted-by":"crossref","DOI":"10.1038\/s41746-018-0061-1","volume":"1","author":"VB Kolachalama","year":"2018","unstructured":"Kolachalama, V. B. & Garg, P. S. Machine learning and medical education. npj Digital Med. 1, 54 (2018).","journal-title":"npj Digital Med."},{"key":"918_CR60","doi-asserted-by":"crossref","first-page":"22071","DOI":"10.1073\/pnas.1900654116","volume":"116","author":"WJ Murdoch","year":"2019","unstructured":"Murdoch, W. J., Singh, C., Kumbier, K., Abbasi-Asl, R. & Yu, B. Definitions, methods, and applications in interpretable machine learning. Proc. Natl Acad. Sci. U.S.A 116, 22071\u201322080 (2019).","journal-title":"Proc. Natl Acad. Sci. U.S.A"},{"key":"918_CR61","doi-asserted-by":"crossref","first-page":"173","DOI":"10.1038\/s41551-018-0324-9","volume":"3","author":"H Lee","year":"2019","unstructured":"Lee, H. et al. An explainable deep-learning algorithm for the detection of acute intracranial haemorrhage from small datasets. Nat. Biomed. Eng. 3, 173\u2013182 (2019).","journal-title":"Nat. Biomed. Eng."}],"container-title":["npj Digital Medicine"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s41746-023-00918-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-023-00918-4","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-023-00918-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,17]],"date-time":"2023-11-17T16:05:40Z","timestamp":1700237140000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s41746-023-00918-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9,14]]},"references-count":61,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2023,12]]}},"alternative-id":["918"],"URL":"https:\/\/doi.org\/10.1038\/s41746-023-00918-4","relation":{},"ISSN":["2398-6352"],"issn-type":[{"value":"2398-6352","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,9,14]]},"assertion":[{"value":"16 June 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 September 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 September 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"N.L. is an Editorial Board Member for <i>npj Digital Medicine<\/i>. They played no role in the peer review of this manuscript. The remaining authors declare that there are no other financial or non-financial competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"172"}}