{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T18:02:41Z","timestamp":1775066561354,"version":"3.50.1"},"reference-count":48,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,4,7]],"date-time":"2025-04-07T00:00:00Z","timestamp":1743984000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,4,7]],"date-time":"2025-04-07T00:00:00Z","timestamp":1743984000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,4,7]]},"DOI":"10.1109\/syscon64521.2025.11014810","type":"proceedings-article","created":{"date-parts":[[2025,5,30]],"date-time":"2025-05-30T17:43:30Z","timestamp":1748627010000},"page":"1-8","source":"Crossref","is-referenced-by-count":10,"title":["Exploring RAG Solutions to Reduce Hallucinations in LLMs"],"prefix":"10.1109","author":[{"given":"Samar","family":"AboulEla","sequence":"first","affiliation":[{"name":"Toronto Metropolitan University,Electrical, Computer, and Biomedical Engineering,Toronto,ON,Canada,M5B 2K3"}]},{"given":"Paria","family":"Zabihitari","sequence":"additional","affiliation":[{"name":"Toronto Metropolitan University,Electrical, Computer, and Biomedical Engineering,Toronto,ON,Canada,M5B 2K3"}]},{"given":"Nourhan","family":"Ibrahim","sequence":"additional","affiliation":[{"name":"Toronto Metropolitan University,Electrical, Computer, and Biomedical Engineering,Toronto,ON,Canada,M5B 2K3"}]},{"given":"Majid","family":"Afshar","sequence":"additional","affiliation":[{"name":"Toronto Metropolitan University,Electrical, Computer, and Biomedical Engineering,Toronto,ON,Canada,M5B 2K3"}]},{"given":"Rasha","family":"Kashef","sequence":"additional","affiliation":[{"name":"Toronto Metropolitan University,Electrical, Computer, and Biomedical Engineering,Toronto,ON,Canada,M5B 2K3"}]}],"member":"263","reference":[{"key":"ref1","article-title":"A survey on medical large language models: Technology, application, trustworthiness, and future directions","author":"Liu","year":"2024","journal-title":"arXiv preprint"},{"key":"ref2","article-title":"A survey on large language models from general purpose to medical applications: Datasets, method-ologies, and evaluations","author":"Wang","year":"2024","journal-title":"arXiv preprint"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1145\/3604237.3626869"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/FIE61694.2024.10892891"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3571730"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671470"},{"key":"ref7","article-title":"Retrieval-augmented generation for large language models: A survey","author":"Gao","year":"2023","journal-title":"arXiv preprint"},{"key":"ref8","article-title":"Rag and rau: A survey on retrieval-augmented language model in natural language processing","author":"Hu","year":"2024","journal-title":"arXiv preprint"},{"key":"ref9","article-title":"Retrieval-augmented generation for natural language processing: A survey","author":"Wu","year":"2024","journal-title":"arXiv preprint"},{"key":"ref10","article-title":"A survey of hallucination in large foundation models","author":"Rawte","year":"2023","journal-title":"arXiv preprint"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.findings-emnlp.27"},{"key":"ref12","first-page":"3929","article-title":"Retrieval augmented language model pretraining","volume-title":"International conference on machine learning","author":"Guu","year":"2020"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.99"},{"key":"ref14","first-page":"9459","article-title":"Retrieval-augmented generation for knowledge-intensive nlp tasks","volume":"33","author":"Lewis","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref15","article-title":"From local to global: A graph rag approach to query-focused summarization","author":"Edge","year":"2024","journal-title":"arXiv preprint"},{"key":"ref16","article-title":"Grag: Graph retrieval-augmented generation","author":"Hu","year":"2024","journal-title":"arXiv preprint"},{"key":"ref17","article-title":"Gnn-rag: Graph neural retrieval for large language model reasoning","author":"Mavromatis","year":"2024","journal-title":"arXiv preprint"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/3571730"},{"key":"ref19","article-title":"Llm hallucinations in practical code generation: Phe-nomena, mechanism, and mitigation","author":"Zhang","year":"2024","journal-title":"arXiv preprint"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671470"},{"key":"ref21","article-title":"Retrieval-augmented generation for large language models: A survey","author":"Gao","year":"2023","journal-title":"arXiv preprint"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.322"},{"key":"ref23","article-title":"Graph retrieval-augmented generation: A survey","author":"Peng","year":"2024","journal-title":"arXiv preprint"},{"key":"ref24","article-title":"Enhancing structured-data retrieval with graphrag: Soccer data case study","author":"Sepasdar","year":"2024","journal-title":"arXiv preprint"},{"key":"ref25","first-page":"132876","article-title":"G-retriever: Retrieval-augmented generation for textual graph understanding and question answering","volume":"37","author":"He","year":"2025","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref26","article-title":"Talk like a graph: Encoding graphs for large language models","author":"Fatemi","year":"2023","journal-title":"arXiv preprint"},{"key":"ref27","article-title":"Graphtext: Graph reasoning in text space","author":"Zhao","year":"2023","journal-title":"arXiv preprint"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/3618260.3649789"},{"key":"ref29","article-title":"Domainrag: A chinese benchmark for evalu-ating domain-specific retrieval-augmented generation","author":"Wang","year":"2024","journal-title":"arXiv preprint"},{"key":"ref30","article-title":"Chain-of-knowledge: Grounding large language models via dynamic knowledge adapting over heterogeneous sources","author":"Li","year":"2024","journal-title":"arXiv preprint"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.matching-1.7"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/tkde.2024.3360454"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00638"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.0601602103"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-019-41695-z"},{"key":"ref36","article-title":"Ragtruth: A hallucination corpus for developing trustworthy retrieval-augmented language models","author":"Wu","year":"2023","journal-title":"arXiv preprint"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i16.29728"},{"key":"ref38","article-title":"What can secondary predictions tell us? an exploration on question-answering with squad-v2. 0","author":"Kamfonas","year":"2022","journal-title":"arXiv preprint"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-86623-4_13"},{"key":"ref40","first-page":"74","article-title":"Rouge: A package for automatic evaluation of sum-maries","volume-title":"Text summarization branches out","author":"Lin","year":"2004"},{"key":"ref41","article-title":"Grapheval: A knowledge-graph based llm hallucination evaluation framework","author":"Sansford","year":"2024","journal-title":"arXiv preprint"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.bionlp-1.13"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-5817"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.3115\/1073083.1073135"},{"key":"ref45","article-title":"Enhancing llm factual accuracy with rag to counter hallucinations: A case study on domain-specific queries in private knowledge-bases","author":"Li","year":"2024","journal-title":"arXiv preprint"},{"key":"ref46","article-title":"Measuring and reducing llm hallucination without gold-standard answers via expertise-weighting","author":"Wei","year":"2024","journal-title":"arXiv preprint"},{"key":"ref47","article-title":"A comprehensive survey of hallucination mitigation techniques in large language models","author":"Tonmoy","year":"2024","journal-title":"arXiv preprint"},{"key":"ref48","article-title":"Distilbert, a distilled version of bert: Smaller, faster, cheaper and lighter. arxiv 2019","author":"Sanh","year":"2019","journal-title":"arXiv preprint"}],"event":{"name":"2025 IEEE International systems Conference (SysCon)","location":"Montreal, QC, Canada","start":{"date-parts":[[2025,4,7]]},"end":{"date-parts":[[2025,4,10]]}},"container-title":["2025 IEEE International systems Conference (SysCon)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11014641\/11014642\/11014810.pdf?arnumber=11014810","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,31]],"date-time":"2025-05-31T04:58:06Z","timestamp":1748667486000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11014810\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,7]]},"references-count":48,"URL":"https:\/\/doi.org\/10.1109\/syscon64521.2025.11014810","relation":{},"subject":[],"published":{"date-parts":[[2025,4,7]]}}}