{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,29]],"date-time":"2026-03-29T09:12:49Z","timestamp":1774775569178,"version":"3.50.1"},"reference-count":57,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2025,4,4]],"date-time":"2025-04-04T00:00:00Z","timestamp":1743724800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,4,4]],"date-time":"2025-04-04T00:00:00Z","timestamp":1743724800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100002808","name":"Carlsbergfondet","doi-asserted-by":"publisher","award":["CF20-0257"],"award-info":[{"award-number":["CF20-0257"]}],"id":[{"id":"10.13039\/501100002808","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001734","name":"Copenhagen University","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100001734","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Ethics Inf Technol"],"published-print":{"date-parts":[[2025,6]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>Machine learning (ML) systems are vulnerable to performance decline over time due to dataset shift. To address this problem, experts often suggest that ML systems should be regularly updated to ensure ongoing performance stability. Some scholarly literature has begun to address the epistemic and ethical challenges associated with different updating methodologies. Thus far, however, little attention has been paid to the impact of model updating on the ML-assisted decision-making process itself. This article aims to address this gap. It argues that model updating introduces a new sub-type of opacity into ML-assisted decision-making\u2014update opacity\u2014that occurs when users cannot understand how or why an update has changed the reasoning or behaviour of an ML system. This type of opacity presents a variety of distinctive epistemic and safety concerns that available solutions to the black box problem in ML are largely ill-equipped to address. A variety of alternative strategies may be developed or pursued to address the problem of update opacity more directly, including bi-factual explanations, dynamic model reporting, and update compatibility. However, each of these strategies presents its own risks or carries significant limitations. Further research will be needed to address the epistemic and safety concerns associated with model updating and update opacity going forward.<\/jats:p>","DOI":"10.1007\/s10676-025-09829-2","type":"journal-article","created":{"date-parts":[[2025,4,5]],"date-time":"2025-04-05T14:01:02Z","timestamp":1743861662000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["A moving target in AI-assisted decision-making: dataset shift, model updating, and the problem of update opacity"],"prefix":"10.1007","volume":"27","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8581-9669","authenticated-orcid":false,"given":"Joshua","family":"Hatherley","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,4]]},"reference":[{"key":"9829_CR1","first-page":"715","volume":"182","author":"GA Adam","year":"2022","unstructured":"Adam, G. A., Chang, C. H. K., Haibe-Kains, B., & Goldenberg, A. (2022). Error amplification when updating deployed machine learning models. Proceedings of Machine Learning Research, 182, 715\u2013740.","journal-title":"Proceedings of Machine Learning Research"},{"key":"9829_CR2","first-page":"1","volume":"31","author":"J Adebayo","year":"2018","unstructured":"Adebayo, J., Gilmer, J., Muelly, M., Goodfellow, I., Hardt, M., & Kim, B. (2018). Sanity checks for saliency maps. Advances in Neural Information Processing Systems, 31, 1\u201311.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"9829_CR3","first-page":"857","volume":"108","author":"B Babic","year":"2023","unstructured":"Babic, B., & Cohen, I. G. (2023). The algorithmic explainability \u201cbait and switch.\u201d Minnesota Law Review, 108, 857\u2013909.","journal-title":"Minnesota Law Review"},{"issue":"6552","key":"9829_CR4","doi-asserted-by":"publisher","first-page":"284","DOI":"10.1126\/science.abg1834","volume":"373","author":"B Babic","year":"2021","unstructured":"Babic, B., Gerke, S., Evgeniou, T., & Cohen, I. G. (2021). Beware explanations from AI in health care. Science, 373(6552), 284\u2013286.","journal-title":"Science"},{"issue":"1","key":"9829_CR5","doi-asserted-by":"publisher","first-page":"2429","DOI":"10.1609\/aaai.v33i01.33012429","volume":"33","author":"G Bansal","year":"2019","unstructured":"Bansal, G., Nushi, B., Kamar, E., Weld, D. S., Lasecki, W. S., & Horvitz, E. (2019). Updates in human-AI teams: Understanding and addressing the performance\/compatibility tradeoff. Proceedings of the AAAI Conference on Artificial Intelligence, 33(1), 2429\u20132437.","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"key":"9829_CR6","doi-asserted-by":"publisher","first-page":"349","DOI":"10.1007\/s13347-019-00391-6","volume":"34","author":"JC Bjerring","year":"2021","unstructured":"Bjerring, J. C., & Busch, J. (2021). Artificial intelligence and patient-centered decision-making. Philosophy and Technology, 34, 349\u2013371.","journal-title":"Philosophy and Technology"},{"key":"9829_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.ejrad.2023.110786","volume":"162","author":"K Borys","year":"2023","unstructured":"Borys, K., Schmitt, Y. A., Nauta, M., Seifert, C., Kr\u00e4mer, N., Friedrich, C. M., & Nensa, F. (2023). Explainable AI in medical imaging: An overview for clinical practitioners\u2014Beyond saliency-based XAI approaches. European Journal of Radiology, 162, 110786.","journal-title":"European Journal of Radiology"},{"issue":"12","key":"9829_CR8","doi-asserted-by":"publisher","DOI":"10.1136\/medethics-2019-105860","volume":"47","author":"M Braun","year":"2021","unstructured":"Braun, M., Hummel, P., Beck, S., & Dabrock, P. (2021). Primer on an ethics of AI-based decision support systems in the clinic. Journal of Medical Ethics, 47(12), e3.","journal-title":"Journal of Medical Ethics"},{"issue":"1","key":"9829_CR9","doi-asserted-by":"publisher","first-page":"205395171562251","DOI":"10.1177\/2053951715622512","volume":"3","author":"J Burrell","year":"2016","unstructured":"Burrell, J. (2016). How the machine \u2018thinks\u2019: Understanding opacity in machine learning algorithms. Big Data & Society, 3(1), 2053951715622512.","journal-title":"Big Data & Society"},{"key":"9829_CR10","doi-asserted-by":"crossref","unstructured":"Cai, C. J., Winter, S., Steiner, D., Wilcox, L., & Terry, M. (2019). \u2018Hello AI\u2019: Uncovering the onboarding needs of medical practitioners for human-AI collaborative decision-making. Proceedings of the ACM on Human\u2013Computer Interaction, 3(CSCW), 1\u201324.","DOI":"10.1145\/3359206"},{"issue":"3","key":"9829_CR11","doi-asserted-by":"publisher","first-page":"231","DOI":"10.1136\/bmjqs-2018-008370","volume":"28","author":"R Challen","year":"2019","unstructured":"Challen, R., Denny, J., Pitt, M., Gompels, L., Edwards, T., & Tsaneva-Atanasova, K. (2019). Artificial intelligence, bias and clinical safety. BMJ Quality and Safety, 28(3), 231\u2013237.","journal-title":"BMJ Quality and Safety"},{"issue":"1","key":"9829_CR12","doi-asserted-by":"publisher","first-page":"271","DOI":"10.1038\/s41746-024-01255-w","volume":"7","author":"J Drogt","year":"2024","unstructured":"Drogt, J., Milota, M., van den Brink, A., & Jongsma, K. (2024). Ethical guidance for reporting and evaluating claims of AI outperforming human doctors. Npj Digital Medicine, 7(1), 271.","journal-title":"Npj Digital Medicine"},{"key":"9829_CR13","doi-asserted-by":"publisher","first-page":"645","DOI":"10.1007\/s11023-018-9481-6","volume":"28","author":"JM Dur\u00e1n","year":"2018","unstructured":"Dur\u00e1n, J. M., & Formanek, N. (2018). Grounds for trust: Essential epistemic opacity and computational reliabilism. Minds and Machines, 28, 645\u2013666.","journal-title":"Minds and Machines"},{"issue":"5","key":"9829_CR14","first-page":"329","volume":"47","author":"JM Dur\u00e1n","year":"2021","unstructured":"Dur\u00e1n, J. M., & Jongsma, K. R. (2021). Who is afraid of black box algorithms? On the epistemological and ethical basis of trust in medical AI. Journal of Medical Ethics, 47(5), 329\u2013335.","journal-title":"Journal of Medical Ethics"},{"key":"9829_CR15","volume":"9","author":"JM Dur\u00e1n","year":"2024","unstructured":"Dur\u00e1n, J. M., van der Vloed, D., Ruifrok, A., & Ypma, R. J. (2024). From understanding to justifying: Computational reliabilism for AI-based forensic evidence evaluation. Forensic Science International: Synergy, 9, 100554.","journal-title":"Forensic Science International: Synergy"},{"issue":"1","key":"9829_CR16","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1038\/s41746-022-00611-y","volume":"5","author":"J Feng","year":"2022","unstructured":"Feng, J., Phillips, R. V., Malenica, I., Bishara, A., Hubbard, A. E., Celi, L. A., & Pirracchio, R. (2022). Clinical artificial intelligence quality improvement: Towards continual monitoring and updating of AI algorithms in healthcare. Npj Digital Medicine, 5(1), 66.","journal-title":"Npj Digital Medicine"},{"issue":"3","key":"9829_CR17","doi-asserted-by":"publisher","first-page":"283","DOI":"10.1056\/NEJMc2104626","volume":"385","author":"SG Finlayson","year":"2021","unstructured":"Finlayson, S. G., Subbaswamy, A., Singh, K., Bowers, J., Kupke, A., Zittrain, J., Kohane, I. S., & Saria, S. (2021). The clinician and dataset shift in artificial intelligence. New England Journal of Medicine, 385(3), 283\u2013286.","journal-title":"New England Journal of Medicine"},{"issue":"2","key":"9829_CR18","doi-asserted-by":"publisher","first-page":"167","DOI":"10.1007\/s11019-022-10076-1","volume":"25","author":"F Funer","year":"2022","unstructured":"Funer, F. (2022). The deception of certainty: How non-interpretable machine learning outcomes challenge the epistemic authority of physicians. A deliberative-relational approach. Medicine, Health Care and Philosophy, 25(2), 167\u2013178.","journal-title":"Medicine, Health Care and Philosophy"},{"issue":"11","key":"9829_CR19","doi-asserted-by":"publisher","first-page":"e745","DOI":"10.1016\/S2589-7500(21)00208-9","volume":"3","author":"M Ghassemi","year":"2021","unstructured":"Ghassemi, M., Oakden-Rayner, L., & Beam, A. L. (2021). The false hope of current approaches to explainable artificial intelligence in health care. The Lancet Digital Health, 3(11), e745\u2013e750.","journal-title":"The Lancet Digital Health"},{"issue":"3","key":"9829_CR20","doi-asserted-by":"publisher","first-page":"205","DOI":"10.1136\/medethics-2019-105586","volume":"46","author":"T Grote","year":"2020","unstructured":"Grote, T., & Berens, P. (2020). On the ethics of algorithmic decision-making in healthcare. Journal of Medical Ethics, 46(3), 205\u2013211.","journal-title":"Journal of Medical Ethics"},{"issue":"5","key":"9829_CR21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3236009","volume":"51","author":"R Guidotti","year":"2018","unstructured":"Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., & Pedreschi, D. (2018). A survey of methods for explaining black box models. ACM Computing Surveys, 51(5), 1\u201342.","journal-title":"ACM Computing Surveys"},{"issue":"4","key":"9829_CR22","doi-asserted-by":"publisher","first-page":"808","DOI":"10.1055\/s-0041-1735184","volume":"12","author":"LL Guo","year":"2021","unstructured":"Guo, L. L., Pfohl, S. R., Fries, J., Posada, J., Fleming, S. L., Aftandilian, C., Shah, N., & Sung, L. (2021). Systematic review of approaches to preserve machine learning performance in the presence of temporal dataset shift in clinical medicine. Applied Clinical Informatics, 12(4), 808\u2013815.","journal-title":"Applied Clinical Informatics"},{"issue":"7","key":"9829_CR23","doi-asserted-by":"publisher","first-page":"478","DOI":"10.1136\/medethics-2019-105935","volume":"46","author":"J Hatherley","year":"2020","unstructured":"Hatherley, J. (2020). Limits of trust in medical AI. Journal of Medical Ethics, 46(7), 478\u2013481.","journal-title":"Journal of Medical Ethics"},{"key":"9829_CR24","doi-asserted-by":"publisher","unstructured":"Hatherley, J. (2024). Data over dialogue: Why artificial intelligence is unlikely to humanise medicine. PhD dissertation, Monash University. https:\/\/doi.org\/10.26180\/24955371.","DOI":"10.26180\/24955371"},{"issue":"2","key":"9829_CR25","doi-asserted-by":"publisher","first-page":"361","DOI":"10.1093\/jamia\/ocac218","volume":"30","author":"J Hatherley","year":"2023","unstructured":"Hatherley, J., & Sparrow, R. (2023). Diachronic and synchronic variation in the performance of adaptive machine learning systems: The ethical challenges. Journal of the American Medical Informatics Association, 30(2), 361\u2013366.","journal-title":"Journal of the American Medical Informatics Association"},{"issue":"3","key":"9829_CR26","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1017\/S0963180122000664","volume":"33","author":"J Hatherley","year":"2024","unstructured":"Hatherley, J., Sparrow, R., & Howard, M. (2024). The virtues of interpretable medical artificial intelligence. Cambridge Quarterly of Healthcare Ethics, 33(3), 323\u2013332.","journal-title":"Cambridge Quarterly of Healthcare Ethics"},{"issue":"1","key":"9829_CR27","first-page":"e04619","volume":"11","author":"K Hoernke","year":"2021","unstructured":"Hoernke, K., Djellouli, N., Andrews, L., Lewis-Jackson, S., Manby, L., Martin, S., Vanderslott, S., & Vindrola-Padros, C. (2021). Frontline healthcare workers\u2019 experiences with personal protective equipment during the COVID-19 pandemic in the UK: A rapid qualitative appraisal. British Medical Journal Open, 11(1), e04619.","journal-title":"British Medical Journal Open"},{"key":"9829_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s41512-020-00090-3","volume":"5","author":"DA Jenkins","year":"2021","unstructured":"Jenkins, D. A., Martin, G. P., Sperrin, M., Riley, R. D., Debray, T. P., Collins, G. S., & Peek, N. (2021). Continual updating and monitoring of clinical prediction models: Time for dynamic prediction systems? Diagnostic and Prognostic Research, 5, 1.","journal-title":"Diagnostic and Prognostic Research"},{"key":"9829_CR29","doi-asserted-by":"crossref","unstructured":"Jesus, S., Bel\u00e9m, C., Balayan, V., Bento, J., Saleiro, P., Bizarro, P., & Gama, J. (2021). How can I choose an explainer? An application-grounded evaluation of post-hoc explanations. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (pp. 805\u2013815).","DOI":"10.1145\/3442188.3445941"},{"issue":"3","key":"9829_CR30","first-page":"37","volume":"39","author":"DH Jonassen","year":"1999","unstructured":"Jonassen, D. H., & Henning, P. (1999). Mental models: Knowledge in the head and knowledge in the world. Educational Technology, 39(3), 37\u201342.","journal-title":"Educational Technology"},{"issue":"3","key":"9829_CR31","doi-asserted-by":"publisher","first-page":"487","DOI":"10.1007\/s13347-019-00372-9","volume":"33","author":"M Krishnan","year":"2020","unstructured":"Krishnan, M. (2020). Against interpretability: A critical examination of the interpretability problem in machine learning. Philosophy and Technology, 33(3), 487\u2013502.","journal-title":"Philosophy and Technology"},{"issue":"2133","key":"9829_CR32","doi-asserted-by":"publisher","first-page":"20180084","DOI":"10.1098\/rsta.2018.0084","volume":"376","author":"JA Kroll","year":"2018","unstructured":"Kroll, J. A. (2018). The fallacy of inscrutability. Philosophical Transactions of the Royal Society a: Mathematical, Physical and Engineering Sciences, 376(2133), 20180084.","journal-title":"Philosophical Transactions of the Royal Society a: Mathematical, Physical and Engineering Sciences"},{"issue":"1","key":"9829_CR33","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1002\/hast.973","volume":"49","author":"AJ London","year":"2019","unstructured":"London, A. J. (2019). Artificial intelligence and black-box medical decisions: Accuracy versus explainability. Hastings Center Report, 49(1), 15\u201321.","journal-title":"Hastings Center Report"},{"key":"9829_CR34","unstructured":"McKernan, B., & Davies, H. (2024). \u2018The machine did it coldly\u2019: Israel used AI to identify 37,000 Hamas targets. The Guardian, April 3rd. https:\/\/www.theguardian.com\/world\/2024\/apr\/03\/israel-gaza-ai-database-hamas-airstrikes."},{"key":"9829_CR35","doi-asserted-by":"publisher","DOI":"10.1017\/S0269888921000102","volume":"36","author":"T Miller","year":"2021","unstructured":"Miller, T. (2021). Contrastive explanation: A structural-model approach. The Knowledge Engineering Review, 36, e14.","journal-title":"The Knowledge Engineering Review"},{"key":"9829_CR36","doi-asserted-by":"crossref","unstructured":"Mitchell, M., Wu, S., Zaldivar, A., Barnes, P., Vasserman, L., Hutchinson, B., Spitzer, E., Raji, I. D., & Gebru, T. (2019). Model cards for model reporting. In Proceedings of the conference on fairness, accountability, and transparency (pp. 220\u2013229). Association for Computing Machinery.","DOI":"10.1145\/3287560.3287596"},{"issue":"5","key":"9829_CR37","doi-asserted-by":"publisher","first-page":"3503","DOI":"10.1007\/s10462-021-10088-y","volume":"55","author":"D Minh","year":"2022","unstructured":"Minh, D., Wang, H. X., Li, Y. F., & Nguyen, T. N. (2022). Explainable artificial intelligence: A comprehensive review. Artificial Intelligence Review, 55(5), 3503\u20133568.","journal-title":"Artificial Intelligence Review"},{"key":"9829_CR38","unstructured":"Molnar, C. (2024). Interpretable machine learning (2nd Edition). https:\/\/christophm.github.io\/interpretable-ml-book\/."},{"issue":"1","key":"9829_CR39","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1207\/s15327108ijap0801_3","volume":"8","author":"K Mosier","year":"1998","unstructured":"Mosier, K., Skitka, L. J., Heers, S., & Burdick, M. (1998). Automation bias: Decision making and performance in high-tech cockpits. The International Journal of Aviation Psychology, 8(1), 47\u201363.","journal-title":"The International Journal of Aviation Psychology"},{"issue":"1","key":"9829_CR40","doi-asserted-by":"publisher","first-page":"16","DOI":"10.1007\/s10676-024-09754-w","volume":"26","author":"A Muralidharan","year":"2024","unstructured":"Muralidharan, A., Savulescu, J., & Schaefer, G. O. (2024). AI and the need for justification (to the patient). Ethics and Information Technology, 26(1), 16.","journal-title":"Ethics and Information Technology"},{"key":"9829_CR41","doi-asserted-by":"crossref","unstructured":"Nguyen, A., Yosinski, J., & Clune, J. (2015). Deep neural networks are easily fooled: High confidence predictions for unrecognizable images. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 427\u2013436).","DOI":"10.1109\/CVPR.2015.7298640"},{"key":"9829_CR42","doi-asserted-by":"crossref","unstructured":"Poursabzi-Sangdeh, F., Goldstein, D. G., Hofman, J. M., Wortman Vaughan, J. W., & Wallach, H. (2021). Manipulating and measuring model interpretability. In Proceedings of the 2021 CHI conference on human factors in computing systems (pp. 1\u201352).","DOI":"10.1145\/3411764.3445315"},{"key":"9829_CR43","doi-asserted-by":"publisher","first-page":"48","DOI":"10.1007\/s10676-023-09721-x","volume":"25","author":"M Pruski","year":"2023","unstructured":"Pruski, M. (2023). Ethics framework for predictive clinical AI model updating. Ethics and Information Technology, 25, 48.","journal-title":"Ethics and Information Technology"},{"key":"9829_CR44","volume-title":"Dataset shift in machine learning","year":"2022","unstructured":"Qui\u00f1onero-Candela, J., Sugiyama, M., Schwaighofer, A., & Lawrence, N. D. (Eds.). (2022). Dataset shift in machine learning. MIT Press."},{"issue":"5","key":"9829_CR45","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1038\/s42256-019-0048-x","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5), 206\u2013215.","journal-title":"Nature Machine Intelligence"},{"issue":"2","key":"9829_CR46","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1162\/99608f92.5a8a3a3d","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin, C., & Radin, J. (2019). Why are we using black box models in AI when we don\u2019t need to? A lesson from an explainable AI competition. Harvard Data Science Review, 1(2), 1\u20139.","journal-title":"Harvard Data Science Review"},{"key":"9829_CR58","doi-asserted-by":"publisher","unstructured":"Shen, J. H., Raji, I. D., & Chen, I Y. (2024). The data addition dilemma. arXiv preprint, 2408.04154 [cs.LG]. https:\/\/doi.org\/10.48550\/arXiv.2408.04154","DOI":"10.48550\/arXiv.2408.04154"},{"issue":"21","key":"9829_CR48","doi-asserted-by":"publisher","first-page":"2199","DOI":"10.1001\/jama.2018.17163","volume":"320","author":"EH Shortliffe","year":"2018","unstructured":"Shortliffe, E. H., & Sep\u00falveda, M. J. (2018). Clinical decision support in the era of artificial intelligence. JAMA, 320(21), 2199\u20132200.","journal-title":"JAMA"},{"issue":"2","key":"9829_CR49","doi-asserted-by":"publisher","DOI":"10.1016\/j.jjimei.2022.100094","volume":"2","author":"V Singh","year":"2022","unstructured":"Singh, V., Chen, S. S., Singhania, M., Nanavati, B., & Gupta, A. (2022). How are reinforcement learning and deep learning algorithms used for big data based decision making in financial industries\u2014A review and research agenda. International Journal of Information Management Data Insights, 2(2), 100094.","journal-title":"International Journal of Information Management Data Insights"},{"issue":"3","key":"9829_CR50","doi-asserted-by":"publisher","first-page":"224","DOI":"10.1017\/can.2024.1","volume":"53","author":"A S\u00f8gaard","year":"2023","unstructured":"S\u00f8gaard, A. (2023). On the opacity of deep neural networks. Canadian Journal of Philosophy, 53(3), 224\u2013239.","journal-title":"Canadian Journal of Philosophy"},{"issue":"2","key":"9829_CR51","doi-asserted-by":"publisher","first-page":"79","DOI":"10.24112\/ijccpm.171678","volume":"17","author":"R Sparrow","year":"2019","unstructured":"Sparrow, R., & Hatherley, J. (2019). The promise and perils of AI in medicine. International Journal of Chinese and Comparative Philosophy of Medicine, 17(2), 79\u2013109.","journal-title":"International Journal of Chinese and Comparative Philosophy of Medicine"},{"issue":"1","key":"9829_CR52","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1002\/hast.1079","volume":"50","author":"R Sparrow","year":"2020","unstructured":"Sparrow, R., & Hatherley, J. (2020). High hopes for \u201cDeep Medicine\u201d? AI, economics, and the future of care. Hastings Center Report, 50(1), 14\u201317.","journal-title":"Hastings Center Report"},{"issue":"10","key":"9829_CR53","doi-asserted-by":"publisher","first-page":"58","DOI":"10.1080\/15265161.2024.2337429","volume":"24","author":"R Sparrow","year":"2024","unstructured":"Sparrow, R., Hatherley, J., Oakley, J., & Bain, C. (2024). Should the use of adaptive machine learning systems in medicine be classified as research? The American Journal of Bioethics, 24(10), 58\u201369.","journal-title":"The American Journal of Bioethics"},{"issue":"2","key":"9829_CR54","first-page":"841","volume":"31","author":"S Wachter","year":"2018","unstructured":"Wachter, S., Mittelstadt, B., & Russell, C. (2018). Counterfactual explanations without opening the black box: Automated decisions and the GDPR. Harvard Journal of Law & Technology, 31(2), 841\u2013888.","journal-title":"Harvard Journal of Law & Technology"},{"issue":"10","key":"9829_CR55","doi-asserted-by":"publisher","first-page":"764","DOI":"10.1136\/medethics-2021-107529","volume":"48","author":"JJ Wadden","year":"2022","unstructured":"Wadden, J. J. (2022). Defining the undefinable: The black box problem in healthcare artificial intelligence. Journal of Medical Ethics, 48(10), 764\u2013768.","journal-title":"Journal of Medical Ethics"},{"issue":"2","key":"9829_CR56","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3489465","volume":"29","author":"AY Wang","year":"2022","unstructured":"Wang, A. Y., Wang, D., Drozdal, J., Muller, M., Park, S., Weisz, J. D., Liu, X., Wu, L., & Dugan, C. (2022). Documentation matters: Human-centered AI system to assist data science code documentation in computational notebooks. ACM Transactions on Computer-Human Interaction, 29(2), 1\u201333.","journal-title":"ACM Transactions on Computer-Human Interaction"},{"issue":"2","key":"9829_CR57","doi-asserted-by":"publisher","first-page":"265","DOI":"10.1007\/s13347-019-00382-7","volume":"34","author":"C Zednik","year":"2021","unstructured":"Zednik, C. (2021). Solving the black box problem: A normative framework for explainable artificial intelligence. Philosophy & Technology, 34(2), 265\u2013288.","journal-title":"Philosophy & Technology"}],"container-title":["Ethics and Information Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10676-025-09829-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10676-025-09829-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10676-025-09829-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T12:17:58Z","timestamp":1757161078000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10676-025-09829-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,4]]},"references-count":57,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,6]]}},"alternative-id":["9829"],"URL":"https:\/\/doi.org\/10.1007\/s10676-025-09829-2","relation":{},"ISSN":["1388-1957","1572-8439"],"issn-type":[{"value":"1388-1957","type":"print"},{"value":"1572-8439","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,4,4]]},"assertion":[{"value":"20 March 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 April 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"20"}}