{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T06:03:06Z","timestamp":1772863386291,"version":"3.50.1"},"reference-count":30,"publisher":"MDPI AG","issue":"5","license":[{"start":{"date-parts":[[2025,4,23]],"date-time":"2025-04-23T00:00:00Z","timestamp":1745366400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["BDCC"],"abstract":"<jats:p>Effective student feedback is fundamental to enhancing learning outcomes in higher education. While traditional assessment methods emphasise both achievements and development areas, the process remains time-intensive for educators. This research explores the application of cognitive computing, specifically open-source Large Language Models (LLMs) Mistral-7B and CodeLlama-7B, to streamline feedback generation for student reports containing both Python programming elements and English narrative content. The findings indicate that these models can provide contextually appropriate feedback on both technical Python coding and English specification and documentation. They effectively identified coding weaknesses and provided constructive suggestions for improvement, as well as insightful feedback on English language quality, structure, and clarity in report writing. These results contribute to the growing body of knowledge on automated assessment feedback in higher education, offering practical insights for institutions considering the implementation of open-source LLMs in their workflows. There are around 22 thousand assessment submissions per year in the School of Computer Science, which is one of eight schools in the Faculty of Engineering and Physical Sciences, which is one of seven faculties in the University of Leeds, which is one of one hundred and sixty-six universities in the UK, so there is clear potential for our methods to scale up to millions of assessment submissions. This study also examines the limitations of current approaches and proposes potential enhancements. The findings support a hybrid system where cognitive computing manages routine tasks and educators focus on complex, personalised evaluations, enhancing feedback quality, consistency, and efficiency in educational settings.<\/jats:p>","DOI":"10.3390\/bdcc9050112","type":"journal-article","created":{"date-parts":[[2025,4,23]],"date-time":"2025-04-23T06:38:37Z","timestamp":1745390317000},"page":"112","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Cognitive Computing with Large Language Models for Student Assessment Feedback"],"prefix":"10.3390","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1434-6497","authenticated-orcid":false,"given":"Noorhan","family":"Abbas","sequence":"first","affiliation":[{"name":"Artificial Intelligence for Language Group, School of Computer Science, University of Leeds, Leeds LS2 9JT, UK"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9395-3764","authenticated-orcid":false,"given":"Eric","family":"Atwell","sequence":"additional","affiliation":[{"name":"Artificial Intelligence for Language Group, School of Computer Science, University of Leeds, Leeds LS2 9JT, UK"}]}],"member":"1968","published-online":{"date-parts":[[2025,4,23]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"70","DOI":"10.3102\/0034654312474350","article-title":"Making sense of assessment feedback in higher education","volume":"83","author":"Evans","year":"2013","journal-title":"Rev. Educ. Res."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"107","DOI":"10.1007\/s11092-013-9187-5","article-title":"Understanding classroom feedback practices: A study of New Zealand student experiences, perceptions, and emotional responses","volume":"26","author":"Harris","year":"2014","journal-title":"Educ. Assess. Eval. Account."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"199","DOI":"10.1080\/03075070600572090","article-title":"Formative Assessment and Self-Regulated Learning: A Model and Seven Principles of Good Feedback Practice","volume":"31","author":"Nicol","year":"2006","journal-title":"Stud. High. Educ."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"1315","DOI":"10.1080\/02602938.2018.1463354","article-title":"The development of student feedback literacy: Enabling uptake of feedback","volume":"43","author":"Carless","year":"2018","journal-title":"Assess. Eval. High. Educ."},{"key":"ref_5","unstructured":"Leiker, D., Finnigan, S., Gyllen, A.R., and Cukurova, M. (2023). Prototyping the use of large language models (llms) for adult learning content creation at scale. arXiv."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"698","DOI":"10.1080\/02602938.2012.691462","article-title":"Rethinking models of feedback for learning: The challenge of design","volume":"38","author":"Boud","year":"2013","journal-title":"Assess. Eval. High. Educ."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"466","DOI":"10.1080\/02602938.2015.1024607","article-title":"Let\u2019s stop the pretence of consistent marking: Exploring the multiple limitations of assessment criteria","volume":"41","author":"Bloxham","year":"2016","journal-title":"Assess. Eval. High. Educ."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"90","DOI":"10.1111\/bjet.13370","article-title":"Practical and ethical challenges of large language models in education: A systematic scoping review","volume":"55","author":"Yan","year":"2024","journal-title":"Br. J. Educ. Technol."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"102274","DOI":"10.1016\/j.lindif.2023.102274","article-title":"ChatGPT for Good? On Opportunities and Challenges of Large Language Models for Education","volume":"103","author":"Kasneci","year":"2023","journal-title":"Learn. Individ. Differ."},{"key":"ref_10","unstructured":"Jiang, A.Q., Sablayrolles, A., Mensch, A., Bamford, C., Chaplot, D.S., Casas, D.D.L., Bressand, F., Lengyel, G., Lample, G., and Saulnier, L. (2023). Mistral 7B. arXiv."},{"key":"ref_11","unstructured":"Roziere, B., Gehring, J., Gloeckle, F., Sootla, S., Gat, I., Tan, X.E., Adi, Y., Liu, J., Sauvestre, R., and Remez, T. (2023). Code llama: Open foundation models for code. arXiv."},{"key":"ref_12","unstructured":"Kluyver, T., Ragan-Kelley, B., P\u00e9rez, F., Granger, B.E., Bussonnier, M., Frederic, J., and Jupyter Development Team (2016). Jupyter Notebooks\u2014A publishing format for reproducible computational workflows. Positioning and Power in Academic Publishing: Players, Agents and Agendas, IOS Press."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Rule, A., Tabard, A., and Hollan, J.D. (2018, January 21\u201326). Exploration and explanation in computational notebooks. Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, Montreal, QC, Canada.","DOI":"10.1145\/3173574.3173606"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"MacNeil, S., Tran, A., Mogil, D., Bernstein, S., Ross, E., and Huang, Z. (2022, January 7\u201311). Generating diverse code explanations using the GPT-3 large language model. Proceedings of the 2022 ACM Conference on International Computing Education Research, Lugano, Switzerland.","DOI":"10.1145\/3501709.3544280"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Viet, T.D., and Markov, K. (2023, January 9\u201311). Using Large Language Models for Bug Localization and Fixing. Proceedings of the 12th International Conference on Awareness Science and Technology (iCAST), Taichung, Taiwan.","DOI":"10.1109\/iCAST57874.2023.10359304"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Savelka, J., Agarwal, A., An, M., Bogart, C., and Sakr, M. (2023, January 7\u201311). Thrilled by your progress! Large language models (GPT-4) no longer struggle to pass assessments in higher education programming courses. Proceedings of the 2023 ACM Conference on International Computing Education Research, Chicago, IL, USA.","DOI":"10.1145\/3568813.3600142"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Dai, W., Lin, J., Jin, H., Li, T., Tsai, Y.-S., Ga\u0161evi\u0107, D., and Chen, G. (2023, January 10\u201313). Can Large Language Models Provide Feedback to Students? ACase Study on ChatGPT. Proceedings of the 2023 IEEE International Conference on Advanced Learning Technologies (ICALT), Orem, UT, USA.","DOI":"10.1109\/ICALT58122.2023.00100"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Yancey, K.P., Laflair, G., Verardi, A., and Burstein, J. (May, January 13). Rating short l2 essays on the CEFR scale with GPT-4. Proceedings of the 18th Workshop on Innovative Use of NLP for Building Educational Applications (BEA 2023), Toronto, ON, Canada.","DOI":"10.18653\/v1\/2023.bea-1.49"},{"key":"ref_19","unstructured":"Xiao, C., Ma, W., Xu, S.X., Zhang, K., Wang, Y., and Fu, Q. (2024). From Automation to Augmentation: Large Language Models Elevating Essay Scoring Landscape. arXiv."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Venter, J., Coetzee, S.A., and Schmulian, A. (2024). Exploring the use of artificial intelligence (AI) in the delivery of effective feedback. Assess. Eval. High. Educ., 1\u201321.","DOI":"10.1080\/02602938.2024.2415649"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Maity, S., and Deroy, A. (2024). Human-Centric eXplainable AI in Education. arXiv.","DOI":"10.35542\/osf.io\/k5u9b"},{"key":"ref_22","unstructured":"Sclar, M., Choi, Y., Tsvetkov, Y., and Suhr, A. (2023). Quantifying Language Models\u2019 Sensitivity to Spurious Features in Prompt Design or: How I learned to start worrying about prompt formatting. arXiv."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Shankar, S., Zamfirescu-Pereira, J., Hartmann, B., Parameswaran, A., and Arawjo, I. (2024, January 13\u201316). Who validates the validators? Aligning LLM-assisted evaluation of LLM outputs with human preferences. Proceedings of the 37th Annual ACM Symposium on User Interface Software and Technology, Pittsburgh, PA, USA.","DOI":"10.1145\/3654777.3676450"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Xu, T., Wu, S., Diao, S., Liu, X., Wang, X., Chen, Y., and Gao, J. (2024). SaySelf: Teaching LLMs to Express Confidence with Self-Reflective Rationales. arXiv.","DOI":"10.18653\/v1\/2024.emnlp-main.343"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"308","DOI":"10.1080\/02601370.2022.2066213","article-title":"Online chat and chatbots to enhance mature student engagement in higher education","volume":"41","author":"Abbas","year":"2022","journal-title":"Int. J. Lifelong Educ."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"e2788","DOI":"10.7717\/peerj-cs.2788","article-title":"Designing AI-powered translation education tools: A framework for parallel sentence generation using SauLTC and LLMs","volume":"11","author":"Aleedy","year":"2025","journal-title":"PeerJ Comput. Sci."},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"100101","DOI":"10.1016\/j.nlp.2024.100101","article-title":"Towards effective teaching assistants: From intent-based chatbots to LLM-powered teaching assistants","volume":"8","author":"Alsafari","year":"2024","journal-title":"Nat. Lang. Process. J."},{"key":"ref_28","unstructured":"Wang, S., Xu, T., Li, H., Zhang, C., Liang, J., Tang, J., Yu, P.S., and Wen, Q. (2024). Large language models for education: A survey and outlook. arXiv."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3571730","article-title":"Survey of hallucination in natural language generation","volume":"55","author":"Ji","year":"2023","journal-title":"ACM Comput. Surv."},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3703155","article-title":"A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions","volume":"43","author":"Huang","year":"2025","journal-title":"ACM Trans. Inf. Syst."}],"container-title":["Big Data and Cognitive Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2504-2289\/9\/5\/112\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T17:19:52Z","timestamp":1760030392000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2504-2289\/9\/5\/112"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,23]]},"references-count":30,"journal-issue":{"issue":"5","published-online":{"date-parts":[[2025,5]]}},"alternative-id":["bdcc9050112"],"URL":"https:\/\/doi.org\/10.3390\/bdcc9050112","relation":{},"ISSN":["2504-2289"],"issn-type":[{"value":"2504-2289","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,4,23]]}}}