{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,17]],"date-time":"2026-04-17T01:49:57Z","timestamp":1776390597122,"version":"3.51.2"},"reference-count":69,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,7,17]],"date-time":"2024-07-17T00:00:00Z","timestamp":1721174400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,7,17]],"date-time":"2024-07-17T00:00:00Z","timestamp":1721174400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Ethics Inf Technol"],"published-print":{"date-parts":[[2024,9]]},"DOI":"10.1007\/s10676-024-09778-2","type":"journal-article","created":{"date-parts":[[2024,7,17]],"date-time":"2024-07-17T05:01:58Z","timestamp":1721192518000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":45,"title":["Beyond transparency and explainability: on the need for adequate and contextualized user guidelines for LLM use"],"prefix":"10.1007","volume":"26","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7277-7351","authenticated-orcid":false,"given":"Kristian Gonz\u00e1lez","family":"Barman","sequence":"first","affiliation":[]},{"given":"Nathan","family":"Wood","sequence":"additional","affiliation":[]},{"given":"Pawel","family":"Pawlowski","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,17]]},"reference":[{"issue":"6","key":"9778_CR1","doi-asserted-by":"publisher","first-page":"461","DOI":"10.1038\/s42256-021-00359-2","volume":"3","author":"A Abid","year":"2021","unstructured":"Abid, A., Farooqi, M., & Zou, J. (2021). Large language models associate Muslims with violence. Nature Machine Intelligence, 3(6), 461\u2013463.","journal-title":"Nature Machine Intelligence"},{"key":"9778_CR2","unstructured":"Agarwal, V., Thureja, N., Garg, M. K., Dharmavaram, S., & Kumar, D. (2024). \u201cWhich LLM should I use?\u201d: Evaluating LLMs for tasks performed by Undergraduate Computer Science Students in India. Preprint retrieved from arXiv:2402.01687."},{"key":"9778_CR3","doi-asserted-by":"crossref","unstructured":"Arrieta, A. B., D\u00edaz-Rodr\u00edguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., ... & Herrera, F. (2020). Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI.\u00a0Information fusion,\u00a058, 82\u2013115.","DOI":"10.1016\/j.inffus.2019.12.012"},{"key":"9778_CR4","unstructured":"Augenstein, I., Baldwin, T., Cha, M., Chakraborty, T., Ciampaglia, G. L., Corney, D., ... & Zagni, G. (2023). Factuality challenges in the era of large language models. Preprint retrieved from arXiv:2310.05189."},{"key":"9778_CR5","doi-asserted-by":"publisher","first-page":"100545","DOI":"10.1016\/j.mlwa.2024.100545","volume":"16","author":"D Barman","year":"2024","unstructured":"Barman, D., Guo, Z., & Conlan, O. (2024). The dark side of language models: Exploring the potential of LLMs in multimedia disinformation generation and dissemination. Machine Learning with Applications, 16, 100545.","journal-title":"Machine Learning with Applications"},{"issue":"1","key":"9778_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11023-024-09657-1","volume":"34","author":"KG Barman","year":"2024","unstructured":"Barman, K. G., Caron, S., Claassen, T., & De Regt, H. (2024b). Towards a benchmark for scientific understanding in humans and machines. Minds and Machines, 34(1), 1\u201316.","journal-title":"Minds and Machines"},{"key":"9778_CR7","doi-asserted-by":"publisher","unstructured":"Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021). On the dangers of stochastic parrots: Can language models be too big? Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (FAccT \u201821), 610\u2013623. https:\/\/doi.org\/10.1145\/3442188.3445922","DOI":"10.1145\/3442188.3445922"},{"key":"9778_CR8","unstructured":"Bills, S., Cammarata, N., Mossing, D., Tillman, H., Gao, L., Goh, G., Sutskever, I., Leike, J., Wu, J., & Saunders, W. (2023) Language models can explain neurons in language models. https:\/\/openaipublic.blob.core.windows.net\/neuron-explainer\/paper\/index.Html"},{"issue":"1","key":"9778_CR9","doi-asserted-by":"publisher","first-page":"43","DOI":"10.1007\/s11023-021-09569-4","volume":"32","author":"FJ Boge","year":"2022","unstructured":"Boge, F. J. (2022). Two dimensions of opacity and the deep learning predicament. Minds and Machines, 32(1), 43\u201375.","journal-title":"Minds and Machines"},{"key":"9778_CR10","unstructured":"Boiko, D. A., MacKnight, R., & Gomes, G. (2023). Emergent autonomous scientific research capabilities of large language models. Preprint retrieved from https:\/\/arxiv.org\/abs\/2304.05332"},{"issue":"1","key":"9778_CR11","doi-asserted-by":"publisher","first-page":"205395171562251","DOI":"10.1177\/2053951715622512","volume":"3","author":"J Burrell","year":"2016","unstructured":"Burrell, J. (2016). How the machine \u2018thinks\u2019: Understanding opacity in machine learning algorithms. Big Data & Society, 3(1), 2053951715622512.","journal-title":"Big Data & Society"},{"key":"9778_CR12","doi-asserted-by":"crossref","unstructured":"Buruk, O\u011fuz\u2019Oz. (2023) Academic Writing with GPT-3.5: Reflections on practices, efficacy and transparency. Preprint retrieved from arXiv:2304.11079.","DOI":"10.31224\/2861"},{"key":"9778_CR13","unstructured":"Chen, C., & Shu, K. (2023). Combating misinformation in the age of LLMs: Opportunities and challenges. Preprint retrieved from arXiv:2311.05656."},{"key":"9778_CR14","doi-asserted-by":"publisher","DOI":"10.56734\/ijahss.v4n11a1","author":"E Choi","year":"2023","unstructured":"Choi, E. (2023). A comprehensive inquiry into the use of ChatGPT: Examining general, educational, and disability-focused perspectives. International Journal of Arts Humanities and Social Sciences. https:\/\/doi.org\/10.56734\/ijahss.v4n11a1","journal-title":"International Journal of Arts Humanities and Social Sciences"},{"key":"9778_CR15","unstructured":"Conmy, A., Mavor-Parker, A. N., Lynch, A., Heimersheim, S., & Garriga-Alonso, A. (2023). Towards automated circuit discovery for mechanistic interpretability. Preprint retrieved from arXiv:2304.14997."},{"key":"9778_CR16","first-page":"16318","volume":"36","author":"A Conmy","year":"2023","unstructured":"Conmy, A., Mavor-Parker, A., Lynch, A., Heimersheim, S., & Garriga-Alonso, A. (2023b). Towards automated circuit discovery for mechanistic interpretability. Advances in Neural Information Processing Systems, 36, 16318\u201316352.","journal-title":"Advances in Neural Information Processing Systems"},{"issue":"1","key":"9778_CR17","first-page":"65","volume":"8","author":"K de Fine Licht","year":"2023","unstructured":"de Fine Licht, K. (2023). Integrating large language models into higher education: guidelines for effective implementation. Computer Sciences & Mathematics Forum, 8(1), 65.","journal-title":"Computer Sciences & Mathematics Forum"},{"issue":"2","key":"9778_CR18","doi-asserted-by":"publisher","first-page":"615","DOI":"10.5114\/biolsport.2023.125623","volume":"40","author":"I Dergaa","year":"2023","unstructured":"Dergaa, I., Chamari, K., Zmijewski, P., & Ben Saad, H. (2023). From human writing to artificial intelligence generated text: Examining the prospects and potential threats of ChatGPT in academic writing. Biology of Sport, 40(2), 615\u2013622. https:\/\/doi.org\/10.5114\/biolsport.2023.125623","journal-title":"Biology of Sport"},{"key":"9778_CR19","doi-asserted-by":"publisher","first-page":"103498","DOI":"10.1016\/j.artint.2021.103498","volume":"297","author":"JM Dur\u00e1n","year":"2021","unstructured":"Dur\u00e1n, J. M. (2021). Dissecting scientific explanation in AI (sXAI): A case for medicine and healthcare. Artificial Intelligence, 297, 103498.","journal-title":"Artificial Intelligence"},{"key":"9778_CR20","doi-asserted-by":"crossref","unstructured":"Eloundou, T., Manning, S., Mishkin, P., & Rock, D. (2023). Gpts are gpts: An early look at the labor market impact potential of large language models. Preprint retrieved from arXiv:2303.10130.","DOI":"10.1126\/science.adj0998"},{"key":"9778_CR21","first-page":"100198","volume":"6","author":"HB Essel","year":"2024","unstructured":"Essel, H. B., Vlachopoulos, D., Essuman, A. B., & Amankwa, J. O. (2024). ChatGPT effects on cognitive skills of undergraduate students: Receiving instant responses from AI-based conversational large language models (LLMs). Computers and Education: Artificial Intelligence, 6, 100198.","journal-title":"Computers and Education: Artificial Intelligence"},{"issue":"7987","key":"9778_CR22","doi-asserted-by":"publisher","first-page":"474","DOI":"10.1038\/d41586-023-03507-3","volume":"623","author":"A Extance","year":"2023","unstructured":"Extance, A. (2023). ChatGPT has entered the classroom: How LLMs could transform education. Nature, 623(7987), 474\u2013477.","journal-title":"Nature"},{"key":"9778_CR23","doi-asserted-by":"publisher","unstructured":"Fan, L., Li, L., Ma, Z., Lee, S., Yu, H., & Hemphill, L. (2023). A bibliometric review of large language models research from 2017 to 2023. Preprint retrieved from https:\/\/doi.org\/10.48550\/arXiv.2304.02020","DOI":"10.48550\/arXiv.2304.02020"},{"issue":"1","key":"9778_CR24","doi-asserted-by":"publisher","first-page":"e51776","DOI":"10.2196\/51776","volume":"6","author":"K Fear","year":"2023","unstructured":"Fear, K., & Gleber, C. (2023). Shaping the future of older adult care: ChatGPT, advanced AI, and the transformation of clinical practice. JMIR Aging, 6(1), e51776.","journal-title":"JMIR Aging"},{"key":"9778_CR25","doi-asserted-by":"crossref","unstructured":"Ferrara, E. (2023). Should chatgpt be biased? Challenges and risks of bias in large language models. Preprint retrieved from arXiv:2304.03738.","DOI":"10.2139\/ssrn.4627814"},{"key":"9778_CR26","doi-asserted-by":"crossref","unstructured":"Gallegos, I. O., Rossi, R. A., Barrow, J., Tanjim, M. M., Kim, S., Dernoncourt, F., ... & Ahmed, N. K. (2023). Bias and fairness in large language models: A survey. Preprint retrieved from arXiv:2309.00770.","DOI":"10.1162\/coli_a_00524"},{"key":"9778_CR27","doi-asserted-by":"crossref","unstructured":"Girotra, K., Meincke, L., Terwiesch, C., & Ulrich, K. T. (2023). Ideas are dimes a dozen: Large language models for idea generation in innovation. Available at SSRN 4526071.","DOI":"10.2139\/ssrn.4526071"},{"issue":"12","key":"9778_CR28","doi-asserted-by":"publisher","first-page":"4876","DOI":"10.1021\/acs.jchemed.3c00505","volume":"100","author":"Y Guo","year":"2023","unstructured":"Guo, Y., & Lee, D. (2023). Leveraging chatgpt for enhancing critical thinking skills. Journal of Chemical Education, 100(12), 4876\u20134883.","journal-title":"Journal of Chemical Education"},{"key":"9778_CR29","doi-asserted-by":"publisher","unstructured":"Hadi, M. U., Al-Tashi, Q., Qureshi, R., Shah, A., Muneer, A., Irfan, M., Zafar, A., Shaikh, M. B., Akhtar, N., Wu, J., Mirjalili, S., & Shah, M. (2023). Large language models: A comprehensive survey of its applications, challenges, limitations, and future prospects. Preprint retrieved from https:\/\/doi.org\/10.36227\/techrxiv.23589741.v4","DOI":"10.36227\/techrxiv.23589741.v4"},{"key":"9778_CR30","doi-asserted-by":"crossref","unstructured":"Hadi, M. U., Qureshi, R., Shah, A., Irfan, M., Zafar, A., Shaikh, M. B., ... & Mirjalili, S. (2023). Large language models: A comprehensive survey of its applications, challenges, limitations, and future prospects. Authorea Preprints.","DOI":"10.36227\/techrxiv.23589741.v2"},{"key":"9778_CR31","doi-asserted-by":"publisher","first-page":"615","DOI":"10.1007\/s11229-008-9435-2","volume":"169","author":"P Humphreys","year":"2009","unstructured":"Humphreys, P. (2009). The philosophical novelty of computer simulation methods. Synthese, 169, 615\u2013626.","journal-title":"Synthese"},{"key":"9778_CR32","doi-asserted-by":"publisher","unstructured":"Inagaki, T., Kato, A., Takahashi, K., Ozaki, H., & Kanda, G. N. (2023). LLMs can generate robotic scripts from goal-oriented instructions in biological laboratory automation. Preprint retrieved from https:\/\/doi.org\/10.48550\/arXiv.2304.10267","DOI":"10.48550\/arXiv.2304.10267"},{"key":"9778_CR33","doi-asserted-by":"publisher","unstructured":"Jablonka, K. M., Ai, Q., Al-Feghali, A., Badhwar, S., Bocarsly, J. D., Bran, A. M., Bringuier, S., Brinson, L. C., Choudhary, K., Circi, D., Cox, S., de Jong, W. A., Evans, M. L., Gastellu, N., Genzling, J., Gil, M. V., Gupta, A. K., Hong, Z., Imran, A., ... Blaiszik, B. (2023). 14 examples of how LLMs can transform materials science and chemistry: A reflection on a large language model hackathon. Digital Discovery, 2(5), 1233\u20131250. https:\/\/doi.org\/10.1039\/d3dd00113j","DOI":"10.1039\/d3dd00113j"},{"key":"9778_CR34","doi-asserted-by":"crossref","unstructured":"Johri, S., Jeong, J., Tran, B. A., Schlessinger, D. I., Wongvibulsin, S., Cai, Z. R., ... & Rajpurkar, P. (2023). Guidelines for rigorous evaluation of clinical LLMs for conversational reasoning. medRxiv, 2023\u201309.","DOI":"10.1101\/2023.09.12.23295399"},{"key":"9778_CR35","doi-asserted-by":"crossref","unstructured":"Kasneci, E., Se\u00dfler, K., K\u00fcchemann, S., Bannert, M., Dementieva, D., Fischer, F., ... & Kasneci, G. (2023). ChatGPT for good? On opportunities and challenges of large language models for education.\u00a0Learning and individual differences,\u00a0103, 102274.","DOI":"10.1016\/j.lindif.2023.102274"},{"key":"9778_CR36","doi-asserted-by":"publisher","first-page":"598","DOI":"10.1016\/j.jpurol.2023.05.018","volume":"19","author":"JK Kim","year":"2023","unstructured":"Kim, J. K., Chua, M., Rickard, M., & Lorenzo, A. (2023). ChatGPT and large language model (LLM) chatbots: The current state of acceptability and a proposal for guidelines on utilization in academic medicine. Journal of Pediatric Urology., 19, 598.","journal-title":"Journal of Pediatric Urology."},{"key":"9778_CR37","doi-asserted-by":"publisher","unstructured":"Lee, J., Le, T., Chen, J., & Lee, D. (2023). Do language models plagiarize? In Proceedings of the ACM Web Conference 2023 (pp. 3637\u20133647). ACM. https:\/\/doi.org\/10.1145\/3543507.3583199","DOI":"10.1145\/3543507.3583199"},{"key":"9778_CR38","unstructured":"Li, Y., Du, M., Song, R., Wang, X., & Wang, Y. (2023). A survey on fairness in large language models. Preprint retrieved from arXiv:2308.10149."},{"key":"9778_CR39","doi-asserted-by":"crossref","unstructured":"Liao, Q. V., & Vaughan, J. W. (2023). Ai transparency in the age of llms: A human-centered research roadmap. Preprint retrieved from arXiv:2306.01941","DOI":"10.1162\/99608f92.8036d03b"},{"issue":"8","key":"9778_CR40","doi-asserted-by":"publisher","first-page":"230658","DOI":"10.1098\/rsos.230658","volume":"10","author":"Z Lin","year":"2023","unstructured":"Lin, Z. (2023). Why and how to embrace AI such as ChatGPT in your academic life. Royal Society Open Science, 10(8), 230658. https:\/\/doi.org\/10.1098\/rsos.230658","journal-title":"Royal Society Open Science"},{"key":"9778_CR41","unstructured":"Lundberg, S. M., & Lee, S. I. (2017). A unified approach to interpreting model predictions. Advances in neural information processing systems, 30."},{"key":"9778_CR42","first-page":"17359","volume":"35","author":"K Meng","year":"2022","unstructured":"Meng, K., Bau, D., Andonian, A., & Belinkov, Y. (2022). Locating and editing factual associations in GPT. Advances in Neural Information Processing Systems, 35, 17359\u201317372.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"9778_CR43","unstructured":"Mishra, A., Soni, U., Arunkumar, A., Huang, J., Kwon, B. C., & Bryan, C. (2023). Promptaid: Prompt exploration, perturbation, testing and iteration using visual analytics for large language models. Preprint retrieved from arXiv:2304.01964."},{"issue":"11","key":"9778_CR44","doi-asserted-by":"publisher","first-page":"1830","DOI":"10.1038\/s41562-023-01744-0","volume":"7","author":"B Mittelstadt","year":"2023","unstructured":"Mittelstadt, B., Wachter, S., & Russell, C. (2023). To protect science, we must use LLMs as zero-shot translators. Nature Human Behaviour, 7(11), 1830\u20131832.","journal-title":"Nature Human Behaviour"},{"key":"9778_CR45","doi-asserted-by":"crossref","unstructured":"Noy, S., & Zhang, W. (2023). Experimental evidence on the productivity effects of generative artificial intelligence. Available at SSRN 4375283.","DOI":"10.2139\/ssrn.4375283"},{"key":"9778_CR46","unstructured":"OpenAI, R. (2023). Gpt-4 technical report. Preprint retrieved from arxiv:2303.08774. View in Article, 2."},{"key":"9778_CR47","doi-asserted-by":"publisher","unstructured":"Pan, Y., Pan, L., Chen, W., Nakov, P., Kan, M.-Y., & Wang, W. Y. (2023). On the risk of misinformation pollution with large language models. Preprint retrieved from https:\/\/doi.org\/10.48550\/arXiv.2305.13661","DOI":"10.48550\/arXiv.2305.13661"},{"key":"9778_CR48","doi-asserted-by":"crossref","unstructured":"Qadir, Junaid. (2023) Engineering education in the era of ChatGPT: Promise and pitfalls of generative AI for education. 2023 IEEE Global Engineering Education Conference (EDUCON). IEEE, 2023.","DOI":"10.1109\/EDUCON54358.2023.10125121"},{"key":"9778_CR49","doi-asserted-by":"publisher","DOI":"10.1177\/01626434231211295","author":"S Rakap","year":"2023","unstructured":"Rakap, S. (2023). Chatting with GPT: Enhancing individualized education program goal development for novice special education teachers. Journal of Special Education Technology. https:\/\/doi.org\/10.1177\/01626434231211295","journal-title":"Journal of Special Education Technology"},{"key":"9778_CR50","unstructured":"Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). Model-agnostic interpretability of machine learning. Preprint retrieved from arXiv:1606.05386."},{"issue":"5","key":"9778_CR51","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1038\/s42256-019-0048-x","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5), 206\u2013215.","journal-title":"Nature Machine Intelligence"},{"key":"9778_CR52","doi-asserted-by":"crossref","unstructured":"Salinas, A., Shah, P., Huang, Y., McCormack, R., & Morstatter, F. (2023, October). The Unequal Opportunities of Large Language Models: Examining Demographic Biases in Job Recommendations by ChatGPT and LLaMA. In\u00a0Proceedings of the 3rd ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization\u00a0(pp. 1\u201315).","DOI":"10.1145\/3617694.3623257"},{"issue":"3","key":"9778_CR53","doi-asserted-by":"publisher","first-page":"258","DOI":"10.1038\/s42256-022-00458-8","volume":"4","author":"P Schramowski","year":"2022","unstructured":"Schramowski, P., Turan, C., Andersen, N., & Herbert, F. (2022). Large pre-trained language models contain human-like biases of what is right and wrong to do. Nature Machine Intelligence, 4(3), 258\u2013268. https:\/\/doi.org\/10.1038\/s42256-022-00458-8","journal-title":"Nature Machine Intelligence"},{"key":"9778_CR54","doi-asserted-by":"publisher","unstructured":"De Silva, D., Mills, N., El-Ayoubi, M., Manic, M., & Alahakoon, D. (2023). ChatGPT and generative AI guidelines for addressing academic integrity and augmenting pre-existing chatbots. In 2023 IEEE International Conference on Industrial Technology (ICIT) (pp. 1\u20136). IEEE. https:\/\/doi.org\/10.1109\/ICIT58465.2023.10143123","DOI":"10.1109\/ICIT58465.2023.10143123"},{"key":"9778_CR55","unstructured":"Sun, Z. (2023). A short survey of viewing large language models in legal aspect. Preprint retrieved from arXiv:2303.09136."},{"key":"9778_CR56","unstructured":"Valentino, M., & Freitas, A. (2022). Scientific explanation and natural language: A unified epistemological-linguistic perspective for explainable AI. Preprint retrieved from arXiv:2205.01809."},{"key":"9778_CR57","doi-asserted-by":"publisher","unstructured":"Vidgof, M., Bachhofner, S., & Mendling, J. (2023). Large language models for business process management: Opportunities and challenges. Preprint retrieved from https:\/\/doi.org\/10.48550\/arXiv.2304.04309","DOI":"10.48550\/arXiv.2304.04309"},{"key":"9778_CR58","unstructured":"Wang, J., Ma, W., Sun, P., Zhang, M., & Nie, J. Y. (2024). Understanding user experience in large language model interactions. Preprint retrieved from arXiv:2401.08329."},{"issue":"1","key":"9778_CR59","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1038\/s41746-024-01029-4","volume":"7","author":"L Wang","year":"2024","unstructured":"Wang, L., Chen, X., Deng, X., Wen, H., You, M., Liu, W., & Li, J. (2024). Prompt engineering in consistency and reliability with the evidence-based guideline for LLMs. npj Digital Medicine, 7(1), 41.","journal-title":"npj Digital Medicine"},{"key":"9778_CR61","doi-asserted-by":"publisher","DOI":"10.1007\/s43681-023-00294-5","author":"R Watkins","year":"2023","unstructured":"Watkins, R. (2023). Guidance for researchers and peer-reviewers on the ethical use of Large Language Models (LLMs) in scientific research workflows. AI and Ethics. https:\/\/doi.org\/10.1007\/s43681-023-00294-5","journal-title":"AI and Ethics"},{"key":"9778_CR62","doi-asserted-by":"publisher","first-page":"24824","DOI":"10.48550\/arXiv.2201.11903","volume":"35","author":"J Wei","year":"2022","unstructured":"Wei, J., Wang, X., Schuurmans, D., Bosma, M., Xia, F., Chi, E., Le, Q. V., & Zhou, D. (2022). Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35, 24824\u201324837. https:\/\/doi.org\/10.48550\/arXiv.2201.11903","journal-title":"Advances in Neural Information Processing Systems"},{"key":"9778_CR63","doi-asserted-by":"publisher","unstructured":"Williams, N., Ivanov, S., & Buhalis, D. (2023). Algorithmic ghost in the research shell: Large language models and academic knowledge creation in management research. Preprint retrieved from https:\/\/doi.org\/10.48550\/arXiv.2303.07304","DOI":"10.48550\/arXiv.2303.07304"},{"issue":"2","key":"9778_CR64","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10676-024-09762-w","volume":"26","author":"NG Wood","year":"2024","unstructured":"Wood, N. G. (2024). Explainable AI in the military domain. Ethics and Information Technology, 26(2), 1\u201313.","journal-title":"Ethics and Information Technology"},{"key":"9778_CR65","doi-asserted-by":"publisher","unstructured":"Xiao, Z., Yuan, X., Liao, Q. V., Abdelghani, R., & Oudeyer, P.-Y. (2023). Supporting qualitative analysis with large language models: Combining codebook with GPT-3 for deductive coding. In Companion Proceedings of the 28th International Conference on Intelligent User Interfaces (pp. 75\u201378). ACM. https:\/\/doi.org\/10.1145\/3581754.3584101","DOI":"10.1145\/3581754.3584101"},{"key":"9778_CR66","doi-asserted-by":"publisher","unstructured":"Yadav, G. (2023). Scaling evidence-based instructional design expertise through large language models. Preprint retrieved from https:\/\/doi.org\/10.48550\/arXiv.2306.01006","DOI":"10.48550\/arXiv.2306.01006"},{"key":"9778_CR67","doi-asserted-by":"publisher","unstructured":"Yan, L., Sha, L., Zhao, L., Li, Y., Martinez-Maldonado, R., Chen, G., Li, X., Jin, Y., & Ga\u0161evi\u0107, D. (2023). Practical and ethical challenges of large language models in education: A systematic literature review. Preprint retrieved from https:\/\/doi.org\/10.48550\/arXiv.2303.13379","DOI":"10.48550\/arXiv.2303.13379"},{"issue":"3","key":"9778_CR68","first-page":"138","volume":"87","author":"MM Yell","year":"2023","unstructured":"Yell, M. M. (2023). Social studies, ChatGPT, and lateral reading. Social Education, 87(3), 138\u2013141.","journal-title":"Social Education"},{"key":"9778_CR69","unstructured":"Zhao, H., Chen, H., Yang, F., Liu, N., Deng, H., Cai, H., ... & Du, M. (2023). Explainability for large language models: A survey. Preprint retrieved from arXiv:2309.01029."},{"key":"9778_CR70","unstructured":"Zolanvari, M., Yang, Z., Khan, K., Jain, R., & Meskin, N. (2021). Trust xai: Model-agnostic explanations for ai with a case study on iiot security. IEEE Internet of Things Journal."}],"updated-by":[{"DOI":"10.1007\/s10676-025-09824-7","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2025,3,10]],"date-time":"2025-03-10T00:00:00Z","timestamp":1741564800000}}],"container-title":["Ethics and Information Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10676-024-09778-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10676-024-09778-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10676-024-09778-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,10]],"date-time":"2025-03-10T08:45:28Z","timestamp":1741596328000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10676-024-09778-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,17]]},"references-count":69,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,9]]}},"alternative-id":["9778"],"URL":"https:\/\/doi.org\/10.1007\/s10676-024-09778-2","relation":{"correction":[{"id-type":"doi","id":"10.1007\/s10676-025-09824-7","asserted-by":"object"}]},"ISSN":["1388-1957","1572-8439"],"issn-type":[{"value":"1388-1957","type":"print"},{"value":"1572-8439","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7,17]]},"assertion":[{"value":"16 May 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 July 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 March 2025","order":3,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Correction","order":4,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"A Correction to this paper has been published:","order":5,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"https:\/\/doi.org\/10.1007\/s10676-025-09824-7","URL":"https:\/\/doi.org\/10.1007\/s10676-025-09824-7","order":6,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflicts of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"47"}}