{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,12]],"date-time":"2025-11-12T13:47:51Z","timestamp":1762955271607,"version":"3.45.0"},"publisher-location":"New York, NY, USA","reference-count":26,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,6,24]]},"DOI":"10.1145\/3721201.3725434","type":"proceedings-article","created":{"date-parts":[[2025,11,12]],"date-time":"2025-11-12T13:35:21Z","timestamp":1762954521000},"page":"464-469","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Extracting Causal Relational Rules for Medical Question-Answering Tasks using Large Language Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-1976-9909","authenticated-orcid":false,"given":"Md Sohanur","family":"Rahman","sequence":"first","affiliation":[{"name":"University of Texas at San Antonio, San Antonio, TX, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3811-3284","authenticated-orcid":false,"given":"Yuexia","family":"Zhang","sequence":"additional","affiliation":[{"name":"University of Texas at San Antonio, San Antonio, TX, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1781-3975","authenticated-orcid":false,"given":"Anthony","family":"Rios","sequence":"additional","affiliation":[{"name":"University of Texas at San Antonio, San Antonio, TX, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1617-5986","authenticated-orcid":false,"given":"Ke","family":"Yang","sequence":"additional","affiliation":[{"name":"University of Texas at San Antonio, San Antonio, TX, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,11,12]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al.","author":"Achiam Josh","year":"2023","unstructured":"Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. 2023. Gpt-4 technical report. arXiv preprint arXiv:2303.08774 (2023)."},{"key":"e_1_3_2_1_2_1","volume-title":"Proceedings of the First Workshop on Patient-Oriented Language Processing (CL4Health) @ LREC-COLING 2024","author":"Andrew Judith Jeyafreeda","year":"2024","unstructured":"Judith Jeyafreeda Andrew, Marc Vincent, Anita Burgun, and Nicolas Garcelon. 2024. Evaluating LLMs for Temporal Entity Extraction from Pediatric Clinical Text in Rare Diseases Context. In Proceedings of the First Workshop on Patient-Oriented Language Processing (CL4Health) @ LREC-COLING 2024, Dina Demner-Fushman, Sophia Ananiadou, Paul Thompson, and Brian Ondov (Eds.). ELRA and ICCL, Torino, Italia, 145\u2013152. https:\/\/aclanthology.org\/2024.cl4health-1.18\/"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-77792-9_6"},{"key":"e_1_3_2_1_4_1","unstructured":"Chunkit Chan Jiayang Cheng Weiqi Wang Yuxin Jiang Tianqing Fang Xin Liu and Yangqiu Song. 2024. ChatGPT Evaluation on Sentence Level Relations: A Focus on Temporal Causal and Discourse Relations. arXiv:2304.14827 [cs.CL] https:\/\/arxiv.org\/abs\/2304.14827"},{"key":"e_1_3_2_1_5_1","first-page":"3936","article-title":"Information Extraction of Aviation Accident Causation Knowledge Graph","volume":"13","author":"Chen Lu","year":"2024","unstructured":"Lu Chen, Jihui Xu, Tianyu Wu, and Jie Liu. 2024. Information Extraction of Aviation Accident Causation Knowledge Graph: An LLM-Based Approach. Electronics 13, 19 (2024), 3936.","journal-title":"An LLM-Based Approach. Electronics"},{"key":"e_1_3_2_1_6_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.3390\/info16010013"},{"key":"e_1_3_2_1_8_1","volume-title":"Proceedings of the 31st International Conference on Computational Linguistics, Owen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa","author":"Hu Zhilei","year":"2025","unstructured":"Zhilei Hu, Zixuan Li, Xiaolong Jin, Long Bai, Jiafeng Guo, and Xueqi Cheng. 2025. Large Language Model-Based Event Relation Extraction with Rationales. In Proceedings of the 31st International Conference on Computational Linguistics, Owen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, and Steven Schockaert (Eds.). Association for Computational Linguistics, Abu Dhabi, UAE, 7484\u20137496. https:\/\/aclanthology.org\/2025.coling-main.500\/"},{"key":"e_1_3_2_1_9_1","volume-title":"First Conference on Language Modeling.","author":"Jiang Haitao","year":"2024","unstructured":"Haitao Jiang, Lin Ge, Yuhe Gao, Jianian Wang, and Rui Song. 2024. LLM4Causal: Democratized Causal Tools for Everyone via Large Language Model. In First Conference on Language Modeling."},{"key":"e_1_3_2_1_10_1","unstructured":"Hyunjae Kim Hyeon Hwang Jiwoo Lee Sihyeon Park Dain Kim Taewhoo Lee Chanwoong Yoon Jiwoong Sohn Donghee Choi and Jaewoo Kang. 2024. Small Language Models Learn Enhanced Reasoning Skills from Medical Textbooks. arXiv:2404.00376 [cs.CL] https:\/\/arxiv.org\/abs\/2404.00376"},{"key":"e_1_3_2_1_11_1","volume-title":"Better zero-shot reasoning with role-play prompting. arXiv preprint arXiv:2308.07702","author":"Kong Aobo","year":"2023","unstructured":"Aobo Kong, Shiwan Zhao, Hao Chen, Qicheng Li, Yong Qin, Ruiqi Sun, Xin Zhou, Enzhi Wang, and Xiaohang Dong. 2023. Better zero-shot reasoning with role-play prompting. arXiv preprint arXiv:2308.07702 (2023)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10115-022-01779-1"},{"key":"e_1_3_2_1_13_1","first-page":"9459","article-title":"Retrieval-augmented generation for knowledge-intensive nlp tasks","volume":"33","author":"Lewis Patrick","year":"2020","unstructured":"Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich K\u00fcttler, Mike Lewis, Wen-tau Yih, Tim Rockt\u00e4schel, et al. 2020. Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems 33 (2020), 9459\u20139474.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2022.110064"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2023.3254132"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"crossref","unstructured":"Sewon Min Xinxi Lyu Ari Holtzman Mikel Artetxe Mike Lewis Hannaneh Hajishirzi and Luke Zettlemoyer. 2022. Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?. In EMNLP.","DOI":"10.18653\/v1\/2022.emnlp-main.759"},{"key":"e_1_3_2_1_17_1","volume-title":"Logesh Kumar Umapathi, and Malaikannan Sankarasubbu","author":"Pal Ankit","year":"2023","unstructured":"Ankit Pal, Logesh Kumar Umapathi, and Malaikannan Sankarasubbu. 2023. Med-HALT: Medical Domain Hallucination Test for Large Language Models. arXiv:2307.15343 [cs.CL] https:\/\/arxiv.org\/abs\/2307.15343"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"crossref","unstructured":"Laria Reynolds and Kyle McDonell. 2021. Prompt programming for large language models: Beyond the few-shot paradigm. In Extended abstracts of the 2021 CHI conference on human factors in computing systems. 1\u20137.","DOI":"10.1145\/3411763.3451760"},{"key":"e_1_3_2_1_19_1","volume-title":"Enhancing Retrieval-Augmented Large Language Models with Iterative Retrieval-Generation Synergy. In The 2023 Conference on Empirical Methods in Natural Language Processing.","author":"Shao Zhihong","year":"2023","unstructured":"Zhihong Shao, Yeyun Gong, Minlie Huang, Nan Duan, Weizhu Chen, et al. 2023. Enhancing Retrieval-Augmented Large Language Models with Iterative Retrieval-Generation Synergy. In The 2023 Conference on Empirical Methods in Natural Language Processing."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.fever-1.20"},{"key":"e_1_3_2_1_21_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yas-mine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale et al. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)."},{"key":"e_1_3_2_1_22_1","volume-title":"AAAI 2024 Workshop on\"Are Large Language Models Simply Causal Parrots?\".","author":"Vashishtha Aniket","year":"2023","unstructured":"Aniket Vashishtha, Abbavaram Gowtham Reddy, Abhinav Kumar, Saketh Bachu, Vineeth N Balasubramanian, and Amit Sharma. 2023. Causal Inference using LLM-Guided Discovery. In AAAI 2024 Workshop on\"Are Large Language Models Simply Causal Parrots?\"."},{"key":"e_1_3_2_1_23_1","volume-title":"Self-Consistency Improves Chain of Thought Reasoning in Language Models. In The Eleventh International Conference on Learning Representations.","author":"Wang Xuezhi","year":"2023","unstructured":"Xuezhi Wang, Jason Wei, Dale Schuurmans, Quoc V Le, Ed H Chi, Sharan Narang, Aakanksha Chowdhery, and Denny Zhou. 2023. Self-Consistency Improves Chain of Thought Reasoning in Language Models. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_24_1","volume-title":"Denny Zhou, et al.","author":"Wei Jason","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le, Denny Zhou, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in neural information processing systems 35 (2022), 24824\u201324837."},{"key":"e_1_3_2_1_25_1","volume-title":"An explanation of in-context learning as implicit bayesian inference. arXiv preprint arXiv:2111.02080","author":"Xie Sang Michael","year":"2021","unstructured":"Sang Michael Xie, Aditi Raghunathan, Percy Liang, and Tengyu Ma. 2021. An explanation of in-context learning as implicit bayesian inference. arXiv preprint arXiv:2111.02080 (2021)."},{"key":"e_1_3_2_1_26_1","volume-title":"Extract","author":"Zhang Bowen","year":"2024","unstructured":"Bowen Zhang and Harold Soh. 2024. Extract, Define, Canonicalize: An LLM-based Framework for Knowledge Graph Construction. arXiv preprint arXiv:2404.03868 (2024)."}],"event":{"name":"CHASE '25: ACM\/IEEE International Conference on Connected Health: Applications, Systems and Engineering Technologies","location":"Yeshiva University Museum New York NY USA","acronym":"CHASE '25","sponsor":["SIGBED ACM Special Interest Group on Embedded Systems","IEEE Computer Society"]},"container-title":["Proceedings of the ACM\/IEEE International Conference on Connected Health: Applications, Systems and Engineering Technologies"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3721201.3725434","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,12]],"date-time":"2025-11-12T13:39:14Z","timestamp":1762954754000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3721201.3725434"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,24]]},"references-count":26,"alternative-id":["10.1145\/3721201.3725434","10.1145\/3721201"],"URL":"https:\/\/doi.org\/10.1145\/3721201.3725434","relation":{},"subject":[],"published":{"date-parts":[[2025,6,24]]},"assertion":[{"value":"2025-11-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}