{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T17:09:36Z","timestamp":1770743376326,"version":"3.49.0"},"reference-count":7,"publisher":"Frontiers Media SA","license":[{"start":{"date-parts":[[2025,4,16]],"date-time":"2025-04-16T00:00:00Z","timestamp":1744761600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["frontiersin.org"],"crossmark-restriction":true},"short-container-title":["Front. Artif. Intell."],"abstract":"<jats:p>Artificial Intelligence (AI) in healthcare holds transformative potential but faces critical challenges in ethical accountability and systemic inequities. Biases in AI models, such as lower diagnosis rates for Black women or gender stereotyping in Large Language Models, highlight the urgent need to address historical and structural inequalities in data and development processes. Disparities in clinical trials and datasets, often skewed toward high-income, English-speaking regions, amplify these issues. Moreover, the underrepresentation of marginalized groups among AI developers and researchers exacerbates these challenges. To ensure equitable AI, diverse data collection, federated data-sharing frameworks, and bias-correction techniques are essential. Structural initiatives, such as fairness audits, transparent AI model development processes, and early registration of clinical AI models, alongside inclusive global collaborations like TRAIN-Europe and CHAI, can drive responsible AI adoption. Prioritizing diversity in datasets and among developers and researchers, as well as implementing transparent governance will foster AI systems that uphold ethical principles and deliver equitable healthcare outcomes globally.<\/jats:p>","DOI":"10.3389\/frai.2025.1577529","type":"journal-article","created":{"date-parts":[[2025,4,16]],"date-time":"2025-04-16T05:27:48Z","timestamp":1744781268000},"update-policy":"https:\/\/doi.org\/10.3389\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["The imperative of diversity and equity for the adoption of responsible AI in healthcare"],"prefix":"10.3389","volume":"8","author":[{"given":"Denise E.","family":"Hilling","sequence":"first","affiliation":[]},{"given":"Imane","family":"Ihaddouchen","sequence":"additional","affiliation":[]},{"given":"Stefan","family":"Buijsman","sequence":"additional","affiliation":[]},{"given":"Reggie","family":"Townsend","sequence":"additional","affiliation":[]},{"given":"Diederik","family":"Gommers","sequence":"additional","affiliation":[]},{"given":"Michel E.","family":"van Genderen","sequence":"additional","affiliation":[]}],"member":"1965","published-online":{"date-parts":[[2025,4,16]]},"reference":[{"key":"ref1","doi-asserted-by":"publisher","first-page":"2232","DOI":"10.1038\/s41591-022-01987-w","article-title":"Tackling bias in AI health datasets through the STANDING together initiative","volume":"28","author":"Ganapathi","year":"2022","journal-title":"Nat. Med."},{"key":"ref2","doi-asserted-by":"publisher","first-page":"e2434997","DOI":"10.1001\/jamanetworkopen.2024.34997","article-title":"Gender representation of health care professionals in large language model\u2013generated stories","volume":"7","author":"Menz","year":"2024","journal-title":"JAMA Netw. Open"},{"key":"ref3","doi-asserted-by":"publisher","first-page":"e213909","DOI":"10.1001\/jamanetworkopen.2021.3909","article-title":"Comparison of methods to reduce bias rom clinical prediction models of postpartum depression","volume":"4","author":"Park","year":"2021","journal-title":"JAMA Netw. Open"},{"key":"ref4","doi-asserted-by":"publisher","first-page":"245","DOI":"10.1001\/jama.2023.26930","article-title":"A nationwide network of health AI assurance laboratories","volume":"331","author":"Shah","year":"","journal-title":"JAMA"},{"key":"ref5","doi-asserted-by":"publisher","first-page":"e2433009","DOI":"10.1001\/jamanetworkopen.2024.33009","article-title":"The need for continuous evaluation of artificial intelligence prediction algorithms","volume":"7","author":"Shah","year":"","journal-title":"JAMA Netw. Open"},{"key":"ref6","doi-asserted-by":"publisher","first-page":"974","DOI":"10.1007\/s00134-024-07408-5","article-title":"Federated data access and federated learning: improved data sharing, AI model development, and learning in intensive care","volume":"50","author":"van Genderen","year":"2024","journal-title":"Intensive Care Med."},{"key":"ref7","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1056\/AIra2400012","article-title":"Clinical text datasets for medical artificial intelligence and large language models\u2014a systematic review","volume":"1","author":"Wu","year":"2024","journal-title":"NEJM AI"}],"container-title":["Frontiers in Artificial Intelligence"],"original-title":[],"link":[{"URL":"https:\/\/www.frontiersin.org\/articles\/10.3389\/frai.2025.1577529\/full","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,16]],"date-time":"2025-04-16T05:27:48Z","timestamp":1744781268000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.frontiersin.org\/articles\/10.3389\/frai.2025.1577529\/full"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,16]]},"references-count":7,"alternative-id":["10.3389\/frai.2025.1577529"],"URL":"https:\/\/doi.org\/10.3389\/frai.2025.1577529","relation":{},"ISSN":["2624-8212"],"issn-type":[{"value":"2624-8212","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,4,16]]},"article-number":"1577529"}}