{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T11:09:05Z","timestamp":1774264145316,"version":"3.50.1"},"reference-count":163,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T00:00:00Z","timestamp":1764806400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T00:00:00Z","timestamp":1764806400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["AI Ethics"],"published-print":{"date-parts":[[2026,2]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>Large language models (LLMs) and their integration into agentic and embodied systems are reshaping artificial intelligence (AI), enabling powerful cross-domain generation and reasoning while introducing new risks. Key concerns include hallucination and misinformation, embedded and amplified biases, privacy leakage, and susceptibility to adversarial manipulation. Ensuring trustworthy and responsible generative AI requires technical reliability, transparency, accountability, and attention to societal impact. The present study conducts a review of peer-reviewed literature on the ethical dimensions of LLMs and LLM-based agents across technical, biomedical, and societal domains. It maps the landscape of risks, distills mitigation strategies (e.g., robust evaluation and red-teaming, alignment and guardrailing, privacy-preserving data practices, bias measurement and reduction, and safety-aware deployment), and examines governance frameworks and operational practices relevant to real-world use. By organizing findings through interdisciplinary lenses and bioethical principles, the review identifies persistent gaps, such as limited context-aware evaluation, uneven reporting standards, and weak post-deployment monitoring, that impede accountability and fairness. The synthesis supports practitioners and policymakers in designing safer, more equitable, and auditable LLM systems, and outlines priorities for future research and governance.<\/jats:p>","DOI":"10.1007\/s43681-025-00847-w","type":"journal-article","created":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T11:07:51Z","timestamp":1764846471000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Ethical perspectives on deployment of large language model agents in biomedicine: a survey"],"prefix":"10.1007","volume":"6","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0068-1531","authenticated-orcid":false,"given":"Nafiseh","family":"Ghaffar Nia","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4257-1496","authenticated-orcid":false,"given":"Amin","family":"Amiri","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0195-7456","authenticated-orcid":false,"given":"Yuan","family":"Luo","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0052-0685","authenticated-orcid":false,"given":"Adrienne","family":"Kline","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,4]]},"reference":[{"key":"847_CR1","unstructured":"Kaur, P., Kashyap, G.S., Kumar, A., Nafis, M.T., Kumar, S., Shokeen, V.: From text to transformation: a comprehensive review of large language models\u2019 versatility, arXiv Preprint arXiv:2402.16142 (2024). Available: https:\/\/arxiv.org\/abs\/2402.16142"},{"key":"847_CR2","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Polosukhin, I.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017). Available: https:\/\/papers.nips.cc\/paperfiles\/paper\/2017\/file\/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf"},{"key":"847_CR3","unstructured":"Jeong, C.: A study on the implementation of generative AI services using an enterprise data-based LLM application architecture, arXiv Preprint arXiv:2309.01105 (2023). Available: https:\/\/arxiv.org\/abs\/2309.01105"},{"issue":"2","key":"847_CR4","volume":"2","author":"DQ Wang","year":"2023","unstructured":"Wang, D.Q., Feng, L.Y., Ye, J.G., Zou, J.G., Zheng, Y.F.: Accelerating the integration of chatgpt and other large-scale AI models into biomedical research and healthcare. MedComm Future Med. 2(2), e43 (2023)","journal-title":"MedComm Future Med."},{"issue":"1","key":"847_CR5","doi-asserted-by":"crossref","first-page":"76","DOI":"10.60087\/jklst.vol1.n1.p84","volume":"1","author":"M Muthusubramanian","year":"2023","unstructured":"Muthusubramanian, M., Jeyaraman, J.: Data engineering innovations: exploring the intersection with cloud computing, machine learning, and AI. J. Knowl. Learn. Sci. Technol. 1(1), 76\u201384 (2023)","journal-title":"J. Knowl. Learn. Sci. Technol."},{"key":"847_CR6","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"240","key":"847_CR7","first-page":"1","volume":"24","author":"A Chowdhery","year":"2023","unstructured":"Chowdhery, A., Narang, S., Devlin, J., Bosma, M., Mishra, G., Roberts, A., Barham, P., Chung, H.W., Sutton, C., Gehrmann, S., et al.: Palm: scaling language modeling with pathways. J. Mach. Learn. Res. 24(240), 1\u2013113 (2023)","journal-title":"J. Mach. Learn. Res."},{"issue":"4s","key":"847_CR8","first-page":"309","volume":"12","author":"DD Patil","year":"2024","unstructured":"Patil, D.D., Dhotre, D.R., Gawande, G.S., Mate, D.S., Shelke, M.V., Bhoye, T.S.: Transformative trends in generative AI: harnessing large language models for natural language understanding and generation. Int. J. Intell. Syst. Appl. Eng. 12(4s), 309\u2013319 (2024)","journal-title":"Int. J. Intell. Syst. Appl. Eng."},{"key":"847_CR9","unstructured":"Achiam, J., et al.: Gpt-4 technical report, arXiv Preprint arXiv:2303.08774 (2023). Available: https:\/\/arxiv.org\/abs\/2303.08774"},{"key":"847_CR10","unstructured":"Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M.-A., Lacroix, T., Rozi\u00e8re., B., Goyal, N., Hambro, E., Azhar, F. et al.: Llama: open and efficient foundation language models, arXiv Preprint arXiv:2302.13971 (2023). Available: https:\/\/arxiv.org\/abs\/2302.13971"},{"key":"847_CR11","unstructured":"Du, N., Huang, Y., Dai, A.M., Tong, S., Lepikhin, D., Xu, Y., Krikun, M., Zhou, Y., Yu, A.W., Firat, O., et al.: Glam: Efficient scaling of language models with mixture-of-experts. In: International Conference on Machine Learning. pp. 5547\u20135569. PMLR (2022). Available: https:\/\/proceedings.mlr.press\/v162\/du22c.html"},{"issue":"3","key":"847_CR12","volume":"10","author":"Q Li","year":"2024","unstructured":"Li, Q., Li, L., Li, Y.: Developing chatgpt for biology and medicine: a complete review of biomedical question answering. Biophys. Rep. 10(3), 152 (2024)","journal-title":"Biophys. Rep."},{"key":"847_CR13","doi-asserted-by":"crossref","unstructured":"Chalkidis, I., Fergadiotis, M., Malakasiotis, P., Aletras, N., Androutsopoulos, I.: Legal-bert: the muppets straight out of law school, arXiv Preprint arXiv:2010.02559 (2020). Available: https:\/\/arxiv.org\/abs\/2010.02559","DOI":"10.18653\/v1\/2020.findings-emnlp.261"},{"key":"847_CR14","unstructured":"Augenstein, I., Baldwin, T., Cha, M., Chakraborty, T., Ciampaglia, G.L., Corney, D, et al.: Factuality challenges in the era of large language models, arXiv Preprint arXiv:2310.05189 (2023). Available: https:\/\/arxiv.org\/abs\/2310.05189"},{"key":"847_CR15","unstructured":"Kaplan, J., McCandlish, S., Henighan, T., Brown, T.B., Chess, B., Child, R., Gray, S., Radford, A., Wu, J., Amodei, D.: Scaling laws for neural language models, arXiv Preprint arXiv:2001.08361 (2020). Available: https:\/\/arxiv.org\/abs\/2001.08361"},{"issue":"140","key":"847_CR16","first-page":"1","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., Liu, P.J.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(140), 1\u201367 (2020)","journal-title":"J. Mach. Learn. Res."},{"key":"847_CR17","unstructured":"Brundage, M., Avin, S., Wang, J., Belfield, H., Krueger, G., Hadfield, G., et al.: Toward trustworthy AI development: mechanisms for supporting verifiable claims. arXiv Preprint arXiv:2004.07213 (2020)"},{"key":"847_CR18","doi-asserted-by":"crossref","unstructured":"Bender, E.M., Gebru, T., McMillan-Major, A., Shmitchell, S.: On the dangers of stochastic parrots: Can language models be too big? In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pp. 610\u2013623. (2021)","DOI":"10.1145\/3442188.3445922"},{"key":"847_CR19","unstructured":"Sanh, V., Debut, L., Chaumond, J., Wolf, T.: Distilbert, a distilled version of bert: Smaller, faster, cheaper and lighter, arXiv Preprint arXiv:1910.01108 (2019)"},{"key":"847_CR20","doi-asserted-by":"crossref","unstructured":"Liu, Z., Zhong, A., Li, Y., Yang, L., Ju, C., Wu, Z., Ma, C., Shu, P., Chen, C., Kim, S., et al.: Tailoring large language models to radiology: a preliminary approach to llm adaptation for a highly specialized domain. In: International Workshop on Machine Learning in Medical Imaging, pp. 464\u2013473. Springer (2023)","DOI":"10.1007\/978-3-031-45673-2_46"},{"issue":"9","key":"847_CR21","doi-asserted-by":"crossref","first-page":"389","DOI":"10.1038\/s42256-019-0088-2","volume":"1","author":"A Jobin","year":"2019","unstructured":"Jobin, A., Ienca, M., Vayena, E.: The global landscape of ai ethics guidelines. Nat. Mach. Intell. 1(9), 389\u2013399 (2019)","journal-title":"Nat. Mach. Intell."},{"issue":"4","key":"847_CR22","doi-asserted-by":"crossref","first-page":"689","DOI":"10.1007\/s11023-018-9482-5","volume":"28","author":"L Floridi","year":"2018","unstructured":"Floridi, L., Cowls, J., Beltrametti, M., Chatila, R., Chazerand, P., Dignum, V., Luetge, C., Madelin, R., Pagallo, U., Rossi, F., Schafer, B., Valcke, P., Vayena, E.: Ai4people, an ethical framework for a good ai society: opportunities, risks, principles, and recommendations. Minds Mach. 28(4), 689\u2013707 (2018)","journal-title":"Minds Mach."},{"key":"847_CR23","unstructured":"Bommasani, R., Hudson, D.A., Adeli, E., et al.: On the opportunities and risks of foundation models, arXiv Preprint arXiv:2108.07258 (2021). Available: https:\/\/arxiv.org\/abs\/2108.07258"},{"issue":"11","key":"847_CR24","doi-asserted-by":"crossref","first-page":"501","DOI":"10.1038\/s42256-019-0114-4","volume":"1","author":"B Mittelstadt","year":"2019","unstructured":"Mittelstadt, B.: Principles alone cannot guarantee ethical ai. Nat. Mach. Intell. 1(11), 501\u2013507 (2019)","journal-title":"Nat. Mach. Intell."},{"key":"847_CR25","unstructured":"European Commission, \u201cProposal for a regulation of the european parliament and the council laying down harmonised rules on artificial intelligence (artificial intelligence act) and amending certain union legislative acts,\u201d EUR-Lex-52021PC0206 (2021). Available: https:\/\/eur-lex.europa.eu\/legal-content\/EN\/TXT\/?uri=CELEX%3A52021PC0206"},{"issue":"2","key":"847_CR26","doi-asserted-by":"crossref","first-page":"191","DOI":"10.1109\/TTS.2024.3403681","volume":"5","author":"J Berengueres","year":"2024","unstructured":"Berengueres, J.: How to regulate large language models for responsible AI. IEEE Trans. Technol. Soc. 5(2), 191\u2013197 (2024)","journal-title":"IEEE Trans. Technol. Soc."},{"key":"847_CR27","unstructured":"Xi, Z., Chen, W., Guo, X., He, W., Ding, Y., Hong, B., et al.: The rise and potential of large language model based agents: A survey, arXiv Preprint arXiv:2309.07864 (2023)"},{"key":"847_CR28","unstructured":"Zhang, J.: Should we fear large language models? A structural analysis of the human reasoning system for elucidating LLM capabilities and risks through the lens of Heidegger\u2019s philosophy, arXiv Preprint arXiv:2403.03288 (2024)"},{"key":"847_CR29","unstructured":"Huang, L., Yu, W., Ma, W., Zhong, W., Feng, Z., Wang, H., et al.: A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions, arXiv Preprint arXiv:2311.05232 (2023)"},{"key":"847_CR30","doi-asserted-by":"crossref","unstructured":"Yang, J., Jin, H., Tang, R., Han, X., Feng, Q., Jiang, H., et al.: Harnessing the power of LLMs in practice: A survey on Chatgpt and beyond. In: ACM Transactions on Knowledge Discovery From Data (2023)","DOI":"10.1145\/3649506"},{"issue":"3","key":"847_CR31","first-page":"86","volume":"6","author":"SR Konda","year":"2022","unstructured":"Konda, S.R.: Ethical considerations in the development and deployment of AI-driven software systems. Int. J. Comput. Sci. Technol. 6(3), 86\u2013101 (2022)","journal-title":"Int. J. Comput. Sci. Technol."},{"key":"847_CR32","doi-asserted-by":"crossref","unstructured":"Nikolinakos, N.T.: Ethical principles for trustworthy AI. In: EU Policy and Legal Framework for Artificial Intelligence, Robotics and Related Technologies\u2013The AI Act, pp. 101\u2013166. Springer (2023)","DOI":"10.1007\/978-3-031-27953-9_3"},{"issue":"4","key":"847_CR33","doi-asserted-by":"crossref","first-page":"230","DOI":"10.1136\/svn-2017-000101","volume":"2","author":"F Jiang","year":"2017","unstructured":"Jiang, F., Jiang, Y., Zhi, H., Dong, Y., Li, H., Ma, S., Wang, Y., Dong, Q., Shen, H., Wang, Y.: Artificial intelligence in healthcare: past, present and future. Stroke Vasc. Neurol. 2(4), 230\u2013243 (2017)","journal-title":"Stroke Vasc. Neurol."},{"key":"847_CR34","doi-asserted-by":"crossref","unstructured":"Shahriari, K., Shahriari, M.: IEEE standard review\u2014ethically aligned design: a vision for prioritizing human wellbeing with artificial intelligence and autonomous systems. In: 2017 IEEE Canada International Humanitarian Technology Conference (IHTC), pp. 197\u2013201. IEEE (2017)","DOI":"10.1109\/IHTC.2017.8058187"},{"key":"847_CR35","doi-asserted-by":"crossref","unstructured":"Sankaran, S.: Enhancing trust through standards: a comparative risk-impact framework for aligning iso ai standards with global ethical and regulatory contexts, arXiv Preprint arXiv:2504.16139 (2025). Available: https:\/\/arxiv.org\/abs\/2504.16139","DOI":"10.1109\/ACDSA65407.2025.11166403"},{"issue":"1","key":"847_CR36","doi-asserted-by":"crossref","first-page":"107","DOI":"10.1186\/s12910-023-00988-9","volume":"24","author":"AR Lee","year":"2023","unstructured":"Lee, A.R., Koo, D., Kim, I.K., Lee, E., Kim, H.H., Yoo, S., Kim, J.-H., Choi, E.K., Lee, H.-Y.: Identifying facilitators of and barriers to the adoption of dynamic consent in digital health ecosystems: a scoping review. BMC Med. Ethics 24(1), 107 (2023)","journal-title":"BMC Med. Ethics"},{"issue":"8","key":"847_CR37","doi-asserted-by":"crossref","first-page":"637","DOI":"10.1001\/jama.2024.0555","volume":"331","author":"KE Goodman","year":"2024","unstructured":"Goodman, K.E., Paul, H.Y., Morgan, D.J.: Ai-generated clinical summaries require more than accuracy. JAMA 331(8), 637\u2013638 (2024)","journal-title":"JAMA"},{"key":"847_CR38","doi-asserted-by":"crossref","unstructured":"Ge, S., Zhou, C., Hou, R., Khabsa, M., Wang, Y.-C., Wang, Q., Han, J., Mao, Y.: Mart: Improving LLM safety with multi-round automatic red-teaming, arXiv preprint arXiv:2311.07689 (2023)","DOI":"10.18653\/v1\/2024.naacl-long.107"},{"issue":"5","key":"847_CR39","doi-asserted-by":"crossref","first-page":"e384","DOI":"10.1016\/S2589-7500(22)00003-6","volume":"4","author":"X Liu","year":"2022","unstructured":"Liu, X., Glocker, B., McCradden, M.M., Ghassemi, M., Denniston, A.K., Oakden-Rayner, L.: The medical algorithmic audit. Lancet Digit. Health 4(5), e384\u2013e397 (2022)","journal-title":"Lancet Digit. Health"},{"key":"847_CR40","doi-asserted-by":"crossref","DOI":"10.3389\/frobt.2021.640647","volume":"8","author":"J Van Der Waa","year":"2021","unstructured":"Van Der Waa, J., Verdult, S., Van Den Bosch, K., Van Diggelen, J., Haije, T., Van Der Stigchel, B., Cocu, I.: Moral decision making in human-agent teams: human control and the role of explanations. Front. Robot. AI 8, 640647 (2021)","journal-title":"Front. Robot. AI"},{"key":"847_CR41","unstructured":"Floridi, L., Cowls, J., Van Wynsberghe, A.: Rethinking meaningful human control for AI systems: from principle to practice, AI and Ethics (2024)"},{"key":"847_CR42","unstructured":"de Cerqueira, J.A.S., Agbese, M., Rousi, R., Xi, N., Hamari, J., Abrahamsson, P.: Can we trust AI agents? An experimental study towards trustworthy LLM-based multi-agent systems for ai ethics, arXiv Preprint arXiv:2411.08881 (2024)"},{"key":"847_CR43","doi-asserted-by":"crossref","unstructured":"Lin, S., Hilton, J., Evans, O.: Truthfulqa: Measuring how models mimic human falsehoods, arXiv Preprint arXiv:2109.07958 (2021)","DOI":"10.18653\/v1\/2022.acl-long.229"},{"key":"847_CR44","unstructured":"Gao, Y., Xiong, Y., Gao, X., Jia, K., Pan, J., Bi, Y., Dai, Y., Sun, J., Wang, H., Wang, H.: Retrieval-augmented generation for large language models: A survey, arXiv Preprint arXiv:2312.10997, vol. 2, no. 1 (2023)"},{"key":"847_CR45","first-page":"27730","volume":"35","author":"L Ouyang","year":"2022","unstructured":"Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., et al.: Training language models to follow instructions with human feedback. Adv. Neural. Inf. Process. Syst. 35, 27730\u201327744 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"847_CR46","unstructured":"Raheja, T., Pochhi, N., Curie, F.: Recent advancements in LLM red-teaming: Techniques, defenses, and ethical considerations, arXiv Preprint arXiv:2410.09097 (2024)"},{"key":"847_CR47","doi-asserted-by":"crossref","unstructured":"Beauchamp, T., Childress, J.: Principles of biomedical ethics: marking its fortieth anniversary, pp. 9\u201312 (2019)","DOI":"10.1080\/15265161.2019.1665402"},{"key":"847_CR48","doi-asserted-by":"crossref","first-page":"285","DOI":"10.1016\/j.jbusres.2021.04.070","volume":"133","author":"N Donthu","year":"2021","unstructured":"Donthu, N., Kumar, S., Mukherjee, D., Pandey, N., Lim, W.M.: How to conduct a bibliometric analysis: an overview and guidelines. J. Bus. Res. 133, 285\u2013296 (2021)","journal-title":"J. Bus. Res."},{"key":"847_CR49","doi-asserted-by":"crossref","unstructured":"H\u00e4ndler, T.: Balancing autonomy and alignment: a multi-dimensional taxonomy for autonomous LLM-powered multi-agent architectures, arXiv Preprint arXiv:2310.03659 (2023)","DOI":"10.5220\/0012239100003598"},{"key":"847_CR50","unstructured":"Ferdaus, M.M., Abdelguerfi, M., Ioup, E., Niles, K.N., Pathak, K., Sloan, S.: Towards trustworthy AI: a review of ethical and robust large language models, arXiv Preprint arXiv:2407.13934 (2024)"},{"key":"847_CR51","unstructured":"Hendrycks, D., Burns, C., Mazeika, M., Song, D.X., Steinhardt, J.: Aligning AI with shared human values. In: International Conference on Machine Learning (ICML), pp. 6438\u20136447. PMLR (2020)"},{"key":"847_CR52","doi-asserted-by":"crossref","unstructured":"Zhang, H., Lu, A.X., Abdalla, M., McDermott, M., Ghassemi, M.: Hurtful words: quantifying biases in clinical contextual word embeddings. In: Proceedings of the ACM Conference on Health, In-Ference, and Learning, pp. 110\u2013120. (2020)","DOI":"10.1145\/3368555.3384448"},{"key":"847_CR53","unstructured":"Yogarajan, V., Dobbie, G., Keegan, T.T., Neuwirth, R.J.: Tackling bias in pre-trained language models: Current trends and under-represented societies, arXiv Preprint arXiv:2312.01509 (2023)"},{"key":"847_CR54","unstructured":"Bai, Y., Kadavath, S., Kundu, S., Askell, A., Kernion, J., Jones, A., et al.: Constitutional AI: harmlessness from AI feedback, arXiv Preprint arXiv:2212.08073 (2022)"},{"key":"847_CR55","first-page":"3008","volume":"35","author":"N Stiennon","year":"2022","unstructured":"Stiennon, N., Ouyang, L., Wu, J., Ziegler, D., et al.: Learning to summarize with human feedback. Adv. Neural. Inf. Process. Syst. 35, 3008\u20133021 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"847_CR56","unstructured":"Glaese, A., McAleese, N., Trebacz, M., Aslanides, J., Firoiu, V., Ewalds, T., et al.: Improving alignment of dialogue agents via targeted human judgements, arXiv Preprint arXiv:2209.14375 (2022)"},{"key":"847_CR57","unstructured":"Partnership on AI, Ai incident database, https:\/\/incidentdatabase.ai\/. Accessed: 2025-06-10 (2025)"},{"issue":"1","key":"847_CR58","first-page":"184","volume":"16","author":"Y-J Chen","year":"2024","unstructured":"Chen, Y.-J., Madisetti, V.K.: Information security, ethics, and integrity in llm agent interaction. J. Inf. Secur. 16(1), 184\u2013196 (2024)","journal-title":"J. Inf. Secur."},{"key":"847_CR59","doi-asserted-by":"crossref","unstructured":"Sreedhar, K., Cai, A., Ma, J., Nickerson, J.V., Chilton, L.B.: Simulating cooperative prosocial behavior with multi-agent LLMs: evidence and mechanisms for AI agents to inform policy decisions. In: Proceedings of the 30th International Conference on Intelligent User Interfaces, pp. 1272\u20131286 (2025)","DOI":"10.1145\/3708359.3712149"},{"key":"847_CR60","doi-asserted-by":"crossref","unstructured":"Dyoub, A., Costantini, S., Letteri, I., Lisi, F.A.: A logic-based multi-agent system for ethical monitoring and evaluation of dialogues, arXiv Preprint arXiv:2109.08294 (2021)","DOI":"10.4204\/EPTCS.345.32"},{"key":"847_CR61","unstructured":"Ruan, Y., Dong, H., Wang, A., Pitis, S., Zhou, Y., Ba, J., et al.: Identifying the risks of LLM agents with an LLM-emulated sandbox, arXiv Preprint arXiv:2309.15817 (2023)"},{"key":"847_CR62","unstructured":"Naihin, S., Atkinson, D., Green, M., Hamadi, M., Swift, C., Schonholtz, D., et al.: Testing language model agents safely in the wild, arXiv Preprint arXiv:2311.10538 (2023)"},{"key":"847_CR63","unstructured":"Tian, Y., Yang, X., Zhang, J., Dong, Y., Su, H.: Evil geniuses: Delving into the safety of LLM-based agents, arXiv Preprint arXiv:2311.11855 (2023)"},{"key":"847_CR64","doi-asserted-by":"crossref","unstructured":"Jung, D., Lee, S., Moon, H., Park, C., Lim, H.: Flex: A benchmark for evaluating robustness of fairness in large language models, arXiv Preprint arXiv:2503.19540 (2025)","DOI":"10.18653\/v1\/2025.findings-naacl.199"},{"key":"847_CR65","unstructured":"Yang, X., Chen, J., Luo, J., Fang, Z., Dong, Y., Su, H., Zhu, J.: Mla-trust: Benchmarking trustworthiness of multimodal LLM agents in GUI environments, arXiv Preprint arXiv:2506.01616 (2025)"},{"key":"847_CR66","unstructured":"Kumar, A., Agarwal, C., Srinivas, S., Li, A.J., Feizi, S., Lakkaraju, H.: Certifying LLM safety against adversarial prompting, arXiv Preprint arXiv:2309.02705 (2023)"},{"key":"847_CR67","doi-asserted-by":"crossref","unstructured":"Gabison, G.A., Xian, R.P.: Inherent and emergent liability issues in LLM-based agentic systems: a principal-agent perspective, arXiv Preprint arXiv:2504.03255 (2025)","DOI":"10.2139\/ssrn.5210666"},{"key":"847_CR68","doi-asserted-by":"crossref","DOI":"10.1016\/j.jss.2022.111507","volume":"195","author":"M Vierhauser","year":"2023","unstructured":"Vierhauser, M., Wohlrab, R., Stadler, M., Cleland-Huang, J.: Amon: a domain-specific language and framework for adaptive monitoring of cyber\u2013physical systems. J. Syst. Softw. 195, 111507 (2023)","journal-title":"J. Syst. Softw."},{"key":"847_CR69","unstructured":"Bolukbasi, T., Chang, K.-W., Zou, J.Y., Saligrama, V., Kalai, A.: Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. In: Proceedings of the 30th Conference on Neural Information Processing Systems (NeurIPS), pp. 4349\u20134357 (2016). Available: https:\/\/proceedings.neurips.cc\/paperfiles\/paper\/2016\/hash\/a486cd07e4ac3d270571622f4f316ec5-Abstract.html"},{"issue":"6","key":"847_CR70","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3457607","volume":"54","author":"N Mehrabi","year":"2021","unstructured":"Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., Galstyan, A.: A survey on bias and fairness in machine learning. ACM Comput. Surv. 54(6), 1\u201335 (2021)","journal-title":"ACM Comput. Surv."},{"key":"847_CR71","doi-asserted-by":"crossref","unstructured":"Bender, E.M., Gebru, T., McMillan-Major, A., Shmitchell, S.: On the dangers of stochastic parrots: Can language models be too big? In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (FAccT), pp. 610\u2013623 (2021)","DOI":"10.1145\/3442188.3445922"},{"key":"847_CR72","doi-asserted-by":"crossref","unstructured":"Wang, Q., Madaio, M., Kane, S., Kapania, S., Terry, M., Wilcox, L.: Designing responsible AI: adaptations of UX practice to meet responsible AI challenges. In: Proceedings of the 2023 Chi Conference on Human Factors in Computing Systems (2023)","DOI":"10.1145\/3544548.3581278"},{"key":"847_CR73","unstructured":"Oroy, K., Nick, A.: Fairness and bias detection in large language models: assessing and mitigating unwanted biases (2024)"},{"key":"847_CR74","unstructured":"Maddalozzo, N.: Methods and measures for bias detection in natural language processing: a study on word embeddings and masked models, Ph.D. dissertation, Politecnico di Torino (2023)"},{"key":"847_CR75","unstructured":"Kumar, A., Singh, S., Murty, S.V., Ragupathy, R.: The ethics of interaction: mitigating security threats in LLMs, arXiv Preprint arXiv:2401.12273 (2024)"},{"key":"847_CR76","doi-asserted-by":"crossref","unstructured":"Zheng, W., Yang, A., Lin, N., Zhou, D.: From bias to fairness: The role of domain-specific knowledge and efficient fine-tuning. In: International Conference on Intelligent Computing, pp. 354\u2013365, Springer (2024)","DOI":"10.1007\/978-981-97-5669-8_29"},{"key":"847_CR77","unstructured":"Deka, P., Jurek-Loughrey, A., Padmanabhan, D.: Evidence-based approach to verification of online health-related content, Queen\u2019s University Belfast (2024)"},{"key":"847_CR78","doi-asserted-by":"crossref","unstructured":"Fabbri, A.R., Wu, C.S., Liu, W., Xiong, C.: Qafacteval: Improved qa-based factual consistency evaluation for summarization, arXiv Preprint arXiv:2112.08542 (2021)","DOI":"10.18653\/v1\/2022.naacl-main.187"},{"key":"847_CR79","unstructured":"Laban, P., Kry\u015bci\u0144ski, W., Agarwal, D., Fabbri, A.R., Xiong, C., Joty, S., Wu, C.S.: LLMs as factual reasoners: Insights from existing benchmarks and beyond, arXiv Preprint arXiv:2305.14540 (2023)"},{"key":"847_CR80","unstructured":"Liu, T., Zhang, Y., Zhao, Z., Dong, Y., Meng, G., Chen, K.: Making them ask and answer: Jailbreaking large language models in few queries via disguise and reconstruction. In: 33rd USENIX Security Symposium (USENIX Security 24), pp. 4711\u20134728 (2024)"},{"issue":"7","key":"847_CR81","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3538707","volume":"55","author":"Z Wang","year":"2022","unstructured":"Wang, Z., Ma, J., Wang, X., Hu, J., Qin, Z., Ren, K.: Threats to training: a survey of poisoning attacks and defenses on machine learning systems. ACM Comput. Surv. 55(7), 1\u201336 (2022)","journal-title":"ACM Comput. Surv."},{"key":"847_CR82","unstructured":"Creo, A., Pudasaini, S.: Silverspeak: Evading AI-generated text detectors using homoglyphs, arXiv Preprint arXiv:2406.11239 (2024)"},{"key":"847_CR83","unstructured":"Smith, V., Shamsabadi, A.S., Ashurst, C., Weller, A.: Identifying and mitigating privacy risks stemming from language models: a survey, arXiv Preprint arXiv:2310.01424 (2023)"},{"key":"847_CR84","unstructured":"Zhang, C., Dai, X., Wu, Y., Yang, Q., Wang, Y., Tang, R., Liu, Y.: A survey on multi-turn interaction capabilities of large language models, arXiv Preprint arXiv:2501.09959 (2025)"},{"key":"847_CR85","doi-asserted-by":"crossref","unstructured":"He, K., Mao, R., Lin, Q., Ruan, Y., Lan, X., Feng, M., et al.: A survey of large language models for healthcare: From data, technology, and applications to accountability and ethics, arXiv Preprint arXiv:2310.05694 (2023)","DOI":"10.2139\/ssrn.4809363"},{"issue":"11s","key":"847_CR86","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3523273","volume":"54","author":"H Hu","year":"2022","unstructured":"Hu, H., Salcic, Z., Sun, L., Dobbie, G., Yu, P.S., Zhang, X.: Membership inference attacks on machine learning: a survey. ACM Comput. Surv. 54(11s), 1\u201337 (2022)","journal-title":"ACM Comput. Surv."},{"issue":"11","key":"847_CR87","first-page":"15","volume":"8","author":"UP Liyanage","year":"2023","unstructured":"Liyanage, U.P., Ranaweera, N.D.: Ethical considerations and potential risks in the deployment of large language models in diverse societal contexts. J. Comput. Soc. Dyn. 8(11), 15\u201325 (2023)","journal-title":"J. Comput. Soc. Dyn."},{"key":"847_CR88","first-page":"687","volume":"69","author":"ER Pike","year":"2019","unstructured":"Pike, E.R.: Defending data: toward ethical protections and comprehensive data governance. Emory Law J. 69, 687\u2013720 (2019)","journal-title":"Emory Law J."},{"key":"847_CR89","doi-asserted-by":"publisher","DOI":"10.1016\/j.hcc.2024.100211","author":"Y Yao","year":"2024","unstructured":"Yao, Y., Duan, J., Xu, K., Cai, Y., Sun, Z., Zhang, Y.: A survey on large language model (LLM) security and privacy: the good, the bad, and the ugly. High-Confidence Comput. (2024). https:\/\/doi.org\/10.1016\/j.hcc.2024.100211","journal-title":"High-Confidence Comput."},{"issue":"1","key":"847_CR90","doi-asserted-by":"crossref","first-page":"40","DOI":"10.1007\/s44163-024-00129-0","volume":"4","author":"IH Sarker","year":"2024","unstructured":"Sarker, I.H.: Llm potentiality and awareness: a position paper from the perspective of trustworthy and responsible AI modeling. Discover Artif. Intell. 4(1), 40 (2024)","journal-title":"Discover Artif. Intell."},{"key":"847_CR91","unstructured":"O. R. OpenAI, GPTs are GPTs: An early look at the labor market impact potential of large language models (2023)"},{"key":"847_CR92","doi-asserted-by":"crossref","unstructured":"Rane, N.: Role of ChatGPT and similar generative artificial intelligence (AI) in construction industry, Available at SSRN 4598258 (2023)","DOI":"10.2139\/ssrn.4598258"},{"issue":"9","key":"847_CR93","first-page":"17","volume":"8","author":"K-J Tokayev","year":"2023","unstructured":"Tokayev, K.-J.: Ethical implications of large language models: a multidimensional exploration of societal, economic, and technical concerns. Int. J. Soc. Anal. 8(9), 17\u201333 (2023)","journal-title":"Int. J. Soc. Anal."},{"issue":"2","key":"847_CR94","doi-asserted-by":"crossref","first-page":"87","DOI":"10.3390\/info16020087","volume":"16","author":"S Pahune","year":"2025","unstructured":"Pahune, S., Akhtar, Z.: Transitioning from mlops to llmops: navigating the unique challenges of large language models. Information 16(2), 87 (2025)","journal-title":"Information"},{"key":"847_CR95","doi-asserted-by":"crossref","unstructured":"Goga, A.S.: Recent developments and ethics of artificial intelligence: safeguards of chatgpt4 and bard, PREPRINT (2023)","DOI":"10.20944\/preprints202311.1211.v1"},{"key":"847_CR96","doi-asserted-by":"crossref","unstructured":"Yao, J., Yi, X., Wang, X., Gong, Y., Xie, X.: Value fulcra: Mapping large language models to the multidimensional spectrum of basic human values, arXiv Preprint arXiv:2311.10766 (2023)","DOI":"10.18653\/v1\/2024.naacl-long.486"},{"issue":"11","key":"847_CR97","doi-asserted-by":"crossref","first-page":"1819","DOI":"10.3390\/math13111819","volume":"13","author":"A Amiri","year":"2025","unstructured":"Amiri, A., Ghaffarnia, A., Nia, N.G., Wu, D., Liang, Y.: Harmonizer: a universal signal tokenization framework for multimodal large language models. Mathematics 13(11), 1819 (2025)","journal-title":"Mathematics"},{"key":"847_CR98","doi-asserted-by":"crossref","unstructured":"Usmani, U.A., Happonen, A., Watada, J.: Human-centered artificial intelligence: designing for user empowerment and ethical considerations. In: 2023 5TH International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA) (2023)","DOI":"10.1109\/HORA58378.2023.10156761"},{"key":"847_CR99","unstructured":"Barocas, S., Hardt, M., Narayanan, A.: Fairness in machine learning. In: Fairness and Machine Learning. fairmlbook.org, Chapter 1 (2017)"},{"key":"847_CR100","doi-asserted-by":"crossref","unstructured":"Sheng, E., Chang, K.W., Natarajan, P., Peng, N.: The woman worked as a babysitter: on biases in language generation. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 3407\u20133412 (2019)","DOI":"10.18653\/v1\/D19-1339"},{"key":"847_CR101","doi-asserted-by":"crossref","DOI":"10.18574\/nyu\/9781479833641.001.0001","volume-title":"Algorithms of oppression: how search engines reinforce racism","author":"SU Noble","year":"2018","unstructured":"Noble, S.U.: Algorithms of oppression: how search engines reinforce racism. NYU Press (2018)"},{"key":"847_CR102","volume-title":"Custodians of the internet: platforms, content moderation, and the hidden decisions that shape social media","author":"T Gillespie","year":"2018","unstructured":"Gillespie, T.: Custodians of the internet: platforms, content moderation, and the hidden decisions that shape social media. Yale University Press (2018)"},{"issue":"1","key":"847_CR103","doi-asserted-by":"crossref","first-page":"110","DOI":"10.1145\/3596490","volume":"67","author":"M Du","year":"2023","unstructured":"Du, M., He, F., Zou, N., Tao, D., Hu, X.: Shortcut learning of large language models in natural language understanding. Commun. ACM 67(1), 110\u2013120 (2023)","journal-title":"Commun. ACM"},{"issue":"2","key":"847_CR104","doi-asserted-by":"crossref","first-page":"856","DOI":"10.3390\/app15020856","volume":"15","author":"A Mohasseb","year":"2025","unstructured":"Mohasseb, A., Amer, E., Chiroma, F., Tranchese, A.: Leveraging advanced nlp techniques and data augmentation to enhance online misogyny detection. Appl. Sci. 15(2), 856 (2025)","journal-title":"Appl. Sci."},{"key":"847_CR105","doi-asserted-by":"crossref","unstructured":"Li, Z., Xu, C.: Discover the unknown biased attribute of an image classifier. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14970\u201314979 (2021)","DOI":"10.1109\/ICCV48922.2021.01470"},{"key":"847_CR106","unstructured":"Islam, M.R.: The evolution of alpha in finance: Harnessing human insight and LLM agents, arXiv Preprint arXiv:2505.14727 (2025)"},{"key":"847_CR107","unstructured":"Guo, J., Cai, H.: System prompt poisoning: Persistent attacks on large language models beyond user injection, arXiv Preprint arXiv:2505.06493 (2025). Available: https:\/\/arxiv.org\/abs\/2505.06493"},{"key":"847_CR108","volume-title":"The language of deception: weaponizing next generation AI","author":"J Hutchens","year":"2023","unstructured":"Hutchens, J.: The language of deception: weaponizing next generation AI. John Wiley & Sons (2023)"},{"key":"847_CR109","doi-asserted-by":"crossref","unstructured":"Pasupuleti, R., Vadapalli, R., Mader, C.: Cyber security issues and challenges related to generative AI and ChatGPT. In: 2023 Tenth International Conference on Social Networks Analysis, Management and Security (SNAMS) (2023)","DOI":"10.1109\/SNAMS60348.2023.10375472"},{"issue":"1","key":"847_CR110","doi-asserted-by":"crossref","first-page":"133","DOI":"10.1111\/joca.12111","volume":"51","author":"GR Milne","year":"2017","unstructured":"Milne, G.R., Pettinico, G., Hajjat, F.M., Markos, E.: Information sensitivity typology: mapping the degree and type of risk consumers perceive in personal data sharing. J. Consum. Aff. 51(1), 133\u2013161 (2017)","journal-title":"J. Consum. Aff."},{"key":"847_CR111","unstructured":"\u201cRegulation (EU) 2024\/1689 of the European parliament and of the council of 13 June 2024 laying down harmonised rules on artificial intelligence (artificial intelligence act), 2024, official Journal of the European Union, L (2024). Available: https:\/\/eur-lex.europa.eu\/eli\/reg\/2024\/1689\/oj\/eng"},{"key":"847_CR112","unstructured":"National Institute of Standards and Technology, \u201cArtificial intelligence risk management framework (AI RMF 1.0),\u201d U.S. Department of Commerce, Tech. Rep (2023). Available: https:\/\/www.nist.gov\/itl\/ai-risk-management-framework"},{"key":"847_CR113","unstructured":"Organisation for Economic Co-operation and Development (OECD), \u201cAdvancing accountability in AI: Governing and managing risks throughout the lifecycle for trustworthy AI,\u201d OECD Digital Economy Papers No. 3492, Tech. Rep., February 2023. Available: https:\/\/www.oecd.org\/content\/dam\/oecd\/en\/publications\/reports\/2023\/02\/advancing-accountability-in-ai753bf8c8\/2448f04b-en.pdf"},{"issue":"2","key":"847_CR114","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3631326","volume":"1","author":"M Hort","year":"2024","unstructured":"Hort, M., Chen, Z., Zhang, J.M., Harman, M., Sarro, F.: Bias mitigation for machine learning classifiers: a comprehensive survey. ACM J. Responsible Comput. 1(2), 1\u201352 (2024)","journal-title":"ACM J. Responsible Comput."},{"issue":"18","key":"847_CR115","doi-asserted-by":"crossref","first-page":"10258","DOI":"10.3390\/app131810258","volume":"13","author":"P Chen","year":"2023","unstructured":"Chen, P., Wu, L., Wang, L.: AI fairness in data management and analytics: a review on challenges, methodologies and applications. Appl. Sci. 13(18), 10258 (2023)","journal-title":"Appl. Sci."},{"issue":"3","key":"847_CR116","first-page":"75","volume":"2","author":"AS George","year":"2024","unstructured":"George, A.S., Baskar, T., Pandey, D.: Establishing global AI accountability: training data transparency, copyright, and misinformation. Partners Univ. Innov. Res. Publ. 2(3), 75\u201391 (2024)","journal-title":"Partners Univ. Innov. Res. Publ."},{"key":"847_CR117","doi-asserted-by":"crossref","unstructured":"Smuha, N.A.: Regulation 2024\/1689 of the Eur. parl. and council of June 13, 2024 (EU artificial intelligence act). In: International Legal Materials, pp. 1\u2013148 (2025)","DOI":"10.1017\/ilm.2024.46"},{"key":"847_CR118","unstructured":"ISO\/IEC 38505-1:2017\u2014Information Technology\u2014Governance of IT\u2014 Governance of Data\u2014Part 1: Application of ISO\/IEC 38500 to the Governance of Data, International Organization for Standardization Std (2017). Available: https:\/\/www.iso.org\/standard\/56639.html"},{"key":"847_CR119","unstructured":"ISO\/IEC 24028:2020\u2014Information Technology\u2014Artificial Intelligence\u2014Overview of Trustworthiness in AI, International Organization for Standardization Std (2020). Available: https:\/\/www.iso.org\/standard\/77608.html"},{"key":"847_CR120","unstructured":"ISO\/IEC 23894:2023\u2014Information Technology\u2014Artificial Intelligence\u2014Guidance on Risk Management, International Organization for Standardization Std (2023). Available: https:\/\/www.iso.org\/standard\/77609.html"},{"issue":"1","key":"847_CR121","doi-asserted-by":"crossref","first-page":"205395171984331","DOI":"10.1177\/2053951719843310","volume":"6","author":"F Torabi Asr","year":"2019","unstructured":"Torabi Asr, F., Taboada, M.: Big data and quality data for fake news and misinformation detection. Big Data Soc. 6(1), 2053951719843310 (2019)","journal-title":"Big Data Soc."},{"issue":"6","key":"847_CR122","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3636551","volume":"56","author":"Y Li","year":"2024","unstructured":"Li, Y., Xie, B., Guo, S., Yang, Y., Xiao, B.: A survey of robustness and safety of 2D and 3D deep learning models against adversarial attacks. ACM Comput. Surv. 56(6), 1\u201337 (2024)","journal-title":"ACM Comput. Surv."},{"key":"847_CR123","unstructured":"Jain, N., Schwarzschild, A., Wen, Y., Somepalli, G., Kirchenbauer, J., Chiang, P.Y., Goldblum, M., Saha, A., Geiping, J., Goldstein, T.: Baseline defenses for adversarial attacks against aligned language models, arXiv Preprint arXiv:2309.00614 (2023). Available: https:\/\/arxiv.org\/abs\/2309.00614"},{"key":"847_CR124","doi-asserted-by":"crossref","unstructured":"Zhao, J.C., Bagchi, S., Avestimehr, S., Chan, K.S., Chaterji, S., Dimitriadis, D., Li, J., Li, N., Nourian, A., Roth, H.R.: Federated learning privacy: Attacks, defenses, applications, and policy landscape, a survey, arXiv Preprint arXiv:2405.03636 (2024)","DOI":"10.1145\/3724113"},{"issue":"5","key":"847_CR125","first-page":"458","volume":"2","author":"M Zaheer","year":"2024","unstructured":"Zaheer, M., Azeem, M.H., Afzal, Z., Karim, H., et al.: Critical evaluation of data privacy and security threats in federated learning: issues and challenges related to privacy and security in IoT. Spectr. Eng. Sci. 2(5), 458\u2013479 (2024)","journal-title":"Spectr. Eng. Sci."},{"key":"847_CR126","doi-asserted-by":"crossref","first-page":"619","DOI":"10.1016\/j.future.2020.10.007","volume":"115","author":"V Mothukuri","year":"2021","unstructured":"Mothukuri, V., Parizi, R.M., Pouriyeh, S., Huang, Y., Dehghantanha, A., Srivastava, G.: A survey on security and privacy of federated learning. Future Gener. Comput. Syst. 115, 619\u2013640 (2021)","journal-title":"Future Gener. Comput. Syst."},{"issue":"3","key":"847_CR127","first-page":"210","volume":"5","author":"I Thomas","year":"2020","unstructured":"Thomas, I.: Getting ready for the California consumer privacy act: building on general data protection regulation preparedness. Appl. Mark. Anal. 5(3), 210\u2013222 (2020)","journal-title":"Appl. Mark. Anal."},{"key":"847_CR128","unstructured":"Wang, Y., Pan, Y., Guo, S., Su, Z.: Security of internet of agents: Attacks and countermeasures, arXiv Preprint arXiv:2505.08807 (2025). Available: https:\/\/arxiv.org\/abs\/2505.08807"},{"issue":"2","key":"847_CR129","first-page":"841","volume":"31","author":"S Wachter","year":"2017","unstructured":"Wachter, S., Mittelstadt, B., Russell, C.: Counterfactual explanations without opening the black box: automated decisions and the gdpr. Harv. J. Law Technol. 31(2), 841\u2013887 (2017)","journal-title":"Harv. J. Law Technol."},{"issue":"3","key":"847_CR130","doi-asserted-by":"crossref","first-page":"348","DOI":"10.1097\/ACM.0000000000004963","volume":"98","author":"RG Russell","year":"2023","unstructured":"Russell, R.G., Novak, L.L., Patel, M., Garvey, K.V., Craig, K.J.T., Jackson, G.P., Moore, D., Miller, B.M.: Competencies for the use of artificial intelligence\u2013based tools by health care professionals. Acad. Med. 98(3), 348\u2013356 (2023)","journal-title":"Acad. Med."},{"key":"847_CR131","volume-title":"The atlas of AI: power, politics, and the planetary costs of artificial intelligence","author":"K Crawford","year":"2021","unstructured":"Crawford, K.: The atlas of AI: power, politics, and the planetary costs of artificial intelligence. Yale University Press (2021)"},{"issue":"8","key":"847_CR132","doi-asserted-by":"crossref","first-page":"1930","DOI":"10.1038\/s41591-023-02448-8","volume":"29","author":"AJ Thirunavukarasu","year":"2023","unstructured":"Thirunavukarasu, A.J., Ting, D.S.J., Elangovan, K., Gutierrez, L., Tan, T.F., Ting, D.S.W.: Large language models in medicine. Nat. Med. 29(8), 1930\u20131940 (2023)","journal-title":"Nat. Med."},{"key":"847_CR133","doi-asserted-by":"crossref","first-page":"102274","DOI":"10.1016\/j.lindif.2023.102274","volume":"103","author":"E Kasneci","year":"2023","unstructured":"Kasneci, E., Se\u00dfler, K., K\u00fcchemann, S., Bannert, M., Dementieva, D., Fischer, F., Gasser, U., Groh, G., G\u00fcnnemann, S., H\u00fcllermeier, E., et al.: Chatgpt for good? On opportunities and challenges of large language models for education. Learn. Individ. Differ. 103, 102274 (2023)","journal-title":"Learn. Individ. Differ."},{"issue":"4","key":"847_CR134","doi-asserted-by":"crossref","first-page":"371","DOI":"10.1001\/jamaophthalmol.2023.6917","volume":"142","author":"AS Huang","year":"2024","unstructured":"Huang, A.S., Hirabayashi, K., Barna, L., Parikh, D., Pasquale, L.R.: Assessment of a large language model\u2019s responses to questions and cases about glaucoma and retina management. JAMA Ophthalmol. 142(4), 371\u2013375 (2024)","journal-title":"JAMA Ophthalmol."},{"issue":"3","key":"847_CR135","doi-asserted-by":"crossref","first-page":"31","DOI":"10.1145\/3236386.3241340","volume":"16","author":"ZC Lipton","year":"2018","unstructured":"Lipton, Z.C.: The mythos of model interpretability: in machine learning, the concept of interpretability is both important and slippery. Queue 16(3), 31\u201357 (2018)","journal-title":"Queue"},{"key":"847_CR136","doi-asserted-by":"crossref","unstructured":"Ribeiro, M. T., Singh, S., Guestrin, C.: \u201cWhy Should I trust you?\u201d: Explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135\u20131144. Association for Computing Machinery, New York (2016)","DOI":"10.1145\/2939672.2939778"},{"key":"847_CR137","doi-asserted-by":"crossref","unstructured":"Nohara, Y., Matsumoto, K., Soejima, H., Nakashima, N.: Explanation of machine learning models using improved shapley additive explanation. In: Proceedings of the 10th ACM International Conference on Bioinformatics, Computational Biology and Health Informatics, pp. 546\u2013546. Association for Computing Machinery, New York (2019)","DOI":"10.1145\/3307339.3343255"},{"key":"847_CR138","unstructured":"Eva, E., H\u00e4ndler, T.: Determinants of LLM-assisted decision-making, arXiv Preprint arXiv:2402.17385 (2024)"},{"key":"847_CR139","unstructured":"Buolamwini, J., Gebru, T.: Gender shades: intersectional accuracy disparities in commercial gender classification. In: Proceedings of the Conference on Fairness, Accountability and Transparency, vol. 81, pp. 77\u201391 (2018)"},{"issue":"4","key":"847_CR140","doi-asserted-by":"crossref","first-page":"607","DOI":"10.1108\/EL-05-2017-0114","volume":"36","author":"F Sheikhshoaei","year":"2018","unstructured":"Sheikhshoaei, F., Naghshineh, N., Alidousti, S., Nakhoda, M.: Design of a digital library maturity model (DLMM). Electron. Libr. 36(4), 607\u2013619 (2018)","journal-title":"Electron. Libr."},{"issue":"2","key":"847_CR141","doi-asserted-by":"crossref","first-page":"56","DOI":"10.1145\/2844110","volume":"59","author":"N Diakopoulos","year":"2016","unstructured":"Diakopoulos, N.: Accountability in algorithmic decision making. Commun. ACM 59(2), 56\u201362 (2016)","journal-title":"Commun. ACM"},{"issue":"1","key":"847_CR142","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1038\/sdata.2016.18","volume":"3","author":"MD Wilkinson","year":"2016","unstructured":"Wilkinson, M.D., Dumontier, M., Aalbersberg, I.J., Appleton, G., Axton, M., Baak, A., Blomberg, N., Boiten, J.W., da Silva Santos, L.B., Bourne, P.E., et al.: The fair guiding principles for scientific data management and stewardship. Sci. Data 3(1), 1\u20139 (2016)","journal-title":"Sci. Data"},{"key":"847_CR143","unstructured":"Anderljung, M., Barnhart, J., Leung, J., Korinek, A., O\u2019Keefe, C., Whittlestone, J., et al.: Frontier AI regulation: managing emerging risks to public safety, arXiv Preprint arXiv:2307.03718 (2023)"},{"key":"847_CR144","doi-asserted-by":"crossref","unstructured":"Sapkota, R., Roumeliotis, K.I., Karkee, M.: Ai agents vs. agentic AI: a conceptual taxonomy, applications and challenge, arXiv Preprint arXiv:2505.10468 (2025). Available: https:\/\/arxiv.org\/abs\/2505.10468","DOI":"10.70777\/si.v2i3.15161"},{"key":"847_CR145","doi-asserted-by":"crossref","first-page":"349","DOI":"10.1023\/B:MIND.0000035461.63578.9d","volume":"14","author":"L Floridi","year":"2004","unstructured":"Floridi, L., Sanders, J.W.: On the morality of artificial agents. Minds Mach. 14, 349\u2013379 (2004)","journal-title":"Minds Mach."},{"key":"847_CR146","unstructured":"Liu, Y., Jia, Y., Geng, R., Jia, J., Gong, N.Z.: Formalizing and benchmarking prompt injection attacks and defenses. In: Proceedings of the 33rd USENIX Security Symposium (USENIX Security 24), pp. 1831\u20131847 (2024). Available: https:\/\/www.usenix.org\/conference\/usenixsecurity24\/presentation\/liu-ya"},{"key":"847_CR147","doi-asserted-by":"crossref","unstructured":"Polyzotis, N., Roy, S., Whang, S.E., Zinkevich, M.: Data management challenges in production machine learning. In: Proceedings of the 2017 ACM International Conference on Management of Data, pp. 1723\u20131726 (2017)","DOI":"10.1145\/3035918.3054782"},{"key":"847_CR148","unstructured":"An, B., Zhang, S., Dredze, M.: Rag llms are not safer: a safety analysis of retrieval-augmented generation for large language models, arXiv Preprint arXiv:2504.18041 (2025). Available: https:\/\/arxiv.org\/abs\/2504.18041"},{"issue":"12","key":"847_CR149","doi-asserted-by":"crossref","first-page":"54","DOI":"10.1145\/3381831","volume":"63","author":"R Schwartz","year":"2020","unstructured":"Schwartz, R., Dodge, J., Smith, N.A., Etzioni, O.: Green AI. Commun. ACM 63(12), 54\u201363 (2020)","journal-title":"Commun. ACM"},{"key":"847_CR150","unstructured":"National Institute of Standards and Technology, Artificial intelligence risk management framework (AI RMF 1.0), U.S. Department of Commerce, Tech. Rep. NIST AI 100-1 (2023)"},{"key":"847_CR151","unstructured":"European Union, Regulation (EU) 2024\/1689 of the European parliament and of the council on artificial intelligence (AI act), Official Journal of the European Union, L 1689 (2024)"},{"issue":"1","key":"847_CR152","doi-asserted-by":"crossref","first-page":"19","DOI":"10.1080\/1364557032000119616","volume":"8","author":"H Arksey","year":"2005","unstructured":"Arksey, H., O\u2019Malley, L.: Scoping studies: towards a methodological framework. Int. J. Soc. Res. Methodol. 8(1), 19\u201332 (2005)","journal-title":"Int. J. Soc. Res. Methodol."},{"key":"847_CR153","doi-asserted-by":"crossref","first-page":"69","DOI":"10.1186\/1748-5908-5-69","volume":"5","author":"D Levac","year":"2010","unstructured":"Levac, D., Colquhoun, H., O\u2019Brien, K.K.: Scoping studies: advancing the methodology. Implement. Sci. 5, 69 (2010)","journal-title":"Implement. Sci."},{"issue":"7","key":"847_CR154","doi-asserted-by":"crossref","first-page":"467","DOI":"10.7326\/M18-0850","volume":"169","author":"AC Tricco","year":"2018","unstructured":"Tricco, A.C., Lillie, E., Zarin, W., O\u2019Brien, K.K., et al.: Prisma extension for scoping reviews (prisma-scr): checklist and explanation. Ann. Intern. Med. 169(7), 467\u2013473 (2018)","journal-title":"Ann. Intern. Med."},{"issue":"4","key":"847_CR155","doi-asserted-by":"crossref","first-page":"219","DOI":"10.3163\/1536-5050.103.4.014","volume":"103","author":"CW Belter","year":"2015","unstructured":"Belter, C.W.: Bibliometric indicators: opportunities and limits. J. Med. Libr. Assoc. 103(4), 219\u2013221 (2015)","journal-title":"J. Med. Libr. Assoc."},{"key":"847_CR156","first-page":"55","volume":"4","author":"G Passas","year":"2024","unstructured":"Passas, G., Xu, L., Kitchin, R.: Bibliometrics and its limitations in AI ethics research evaluation. AI Ethics 4, 55\u201371 (2024)","journal-title":"AI Ethics"},{"issue":"1","key":"847_CR157","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s13643-021-01675-9","volume":"10","author":"D Pieper","year":"2021","unstructured":"Pieper, D., Allers, K.: Language restrictions in systematic reviews should not be imposed arbitrarily: a case study. Syst. Rev. 10(1), 1\u20139 (2021)","journal-title":"Syst. Rev."},{"key":"847_CR158","first-page":"93","volume":"125","author":"B Nussbaumer-Streit","year":"2020","unstructured":"Nussbaumer-Streit, B., Klerings, I., Dobrescu, A.I., et al.: Excluding non-english publications from evidence syntheses: empirical analysis of bias in meta-analyses. J. Clin. Epidemiol. 125, 93\u2013101 (2020)","journal-title":"J. Clin. Epidemiol."},{"key":"847_CR159","doi-asserted-by":"crossref","unstructured":"Deng, Y., Zhang, J., Wang, H., Liu, Z.: Investigating data contamination in language model benchmarks. In: Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL) (2024)","DOI":"10.18653\/v1\/2024.naacl-long.482"},{"key":"847_CR160","doi-asserted-by":"crossref","unstructured":"Yao, J., Xu, Y., Tang, Z., Li, Y.: Data contamination in cross-lingual evaluation of large language models. In: Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP) (2024)","DOI":"10.18653\/v1\/2024.emnlp-main.990"},{"key":"847_CR161","first-page":"155","volume":"12","author":"S Kim","year":"2024","unstructured":"Kim, S., Ramesh, A., Lee, H.: Evaluation data contamination: impacts and mitigation strategies for LLM assessment. Trans. Assoc. Comput. Linguist. 12, 155\u2013173 (2024)","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"847_CR162","unstructured":"Huang, J., Liu, Y., Tang, J., Chen, D.: A survey on hallucination in large language models: Principles, taxonomy, and open challenges, arXiv preprint arXiv:2311.05232 (2023)"},{"key":"847_CR163","unstructured":"Casper, S., Hadfield-Menell, D.: Open problems and fundamental limitations of reinforcement learning from human feedback, arXiv preprint arXiv:2307.15217 (2023)"}],"container-title":["AI and Ethics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-025-00847-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s43681-025-00847-w","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-025-00847-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T10:25:59Z","timestamp":1774261559000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s43681-025-00847-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,4]]},"references-count":163,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["847"],"URL":"https:\/\/doi.org\/10.1007\/s43681-025-00847-w","relation":{},"ISSN":["2730-5953","2730-5961"],"issn-type":[{"value":"2730-5953","type":"print"},{"value":"2730-5961","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,4]]},"assertion":[{"value":"13 October 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 November 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"All authors consent to publication of the manuscript.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}],"article-number":"32"}}