{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T06:25:47Z","timestamp":1774506347888,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":54,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,21]],"date-time":"2024-10-21T00:00:00Z","timestamp":1729468800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Hong Kong ITC Innovation and Technology Fund Midstream Research Programme for Universities Project","award":["No.ITS\/034\/22MS"],"award-info":[{"award-number":["No.ITS\/034\/22MS"]}]},{"name":"CityU - HKIDS Early Career Research Grant","award":["No.9360163"],"award-info":[{"award-number":["No.9360163"]}]},{"name":"Tencent Rhino-Bird Focused Research Program"},{"name":"CCF-Tencent Open Fund"},{"name":"APRC - CityU New Research Initiatives","award":["No.9610565, Start-up Grant for New Faculty of CityU"],"award-info":[{"award-number":["No.9610565, Start-up Grant for New Faculty of CityU"]}]},{"name":"Hong Kong Environmental and Conservation Fund","award":["No. 88\/2022"],"award-info":[{"award-number":["No. 88\/2022"]}]},{"name":"Research Impact Fund","award":["No.R1015-23"],"award-info":[{"award-number":["No.R1015-23"]}]},{"name":"SIRG - CityU Strategic Interdisciplinary Research Grant","award":["No.7020046"],"award-info":[{"award-number":["No.7020046"]}]},{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No.62222213, U23A20319, 62072423"],"award-info":[{"award-number":["No.62222213, U23A20319, 62072423"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,21]]},"DOI":"10.1145\/3627673.3679673","type":"proceedings-article","created":{"date-parts":[[2024,10,20]],"date-time":"2024-10-20T19:34:21Z","timestamp":1729452861000},"page":"2660-2670","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":14,"title":["Editing Factual Knowledge and Explanatory Ability of Medical Large Language Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3971-9907","authenticated-orcid":false,"given":"Derong","family":"Xu","sequence":"first","affiliation":[{"name":"University of Science and Technology of China &amp; City University of Hong Kong, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2968-7304","authenticated-orcid":false,"given":"Ziheng","family":"Zhang","sequence":"additional","affiliation":[{"name":"Jarvis Research Center, Tencent YouTu Lab, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-4530-5516","authenticated-orcid":false,"given":"Zhihong","family":"Zhu","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1264-6549","authenticated-orcid":false,"given":"Zhenxi","family":"Lin","sequence":"additional","affiliation":[{"name":"Jarvis Research Center, Tencent YouTu Lab, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0751-2602","authenticated-orcid":false,"given":"Qidong","family":"Liu","sequence":"additional","affiliation":[{"name":"City University of Hong Kong &amp; Xi'an Jiaotong University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1118-9710","authenticated-orcid":false,"given":"Xian","family":"Wu","sequence":"additional","affiliation":[{"name":"Jarvis Research Center, Tencent YouTu Lab, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4246-5386","authenticated-orcid":false,"given":"Tong","family":"Xu","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5976-0707","authenticated-orcid":false,"given":"Wanyu","family":"Wang","sequence":"additional","affiliation":[{"name":"City University of Hong Kong, Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1513-7814","authenticated-orcid":false,"given":"Yuyang","family":"Ye","sequence":"additional","affiliation":[{"name":"City University of Hong Kong, Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2926-4416","authenticated-orcid":false,"given":"Xiangyu","family":"Zhao","sequence":"additional","affiliation":[{"name":"City University of Hong Kong, Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4835-4102","authenticated-orcid":false,"given":"Enhong","family":"Chen","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2195-2847","authenticated-orcid":false,"given":"Yefeng","family":"Zheng","sequence":"additional","affiliation":[{"name":"Jarvis Research Center, Tencent YouTu Lab &amp; Medical Artificial Intelligence Lab, Westlake University, Shenzhen, China"}]}],"member":"320","published-online":{"date-parts":[[2024,10,21]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3641289"},{"key":"e_1_3_2_2_2_1","unstructured":"Zeming Chen Alejandro Hern\u00e1ndez-Cano Angelika Romanou Antoine Bonnet Kyle Matoba Francesco Salvi Matteo Pagliardini Simin Fan Andreas K\u00f6pf Amirkeivan Mohtashami Alexandre Sallinen Alireza Sakhaeirad Vinitra Swamy Igor Krawczuk Deniz Bayazit Axel Marmet Syrielle Montariol Mary-Anne Hartley Martin Jaggi and Antoine Bosselut. 2023. MEDITRON-70B: Scaling Medical Pretraining for Large Language Models. arxiv: 2311.16079 [cs.CL]"},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.naacl-long.73"},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.522"},{"key":"e_1_3_2_2_6_1","volume-title":"Calibrating Factual Knowledge in Pretrained Language Models. arXiv preprint arXiv:2210.03329","author":"Dong Qingxiu","year":"2022","unstructured":"Qingxiu Dong, Damai Dai, Yifan Song, Jingjing Xu, Zhifang Sui, and Lei Li. 2022. Calibrating Factual Knowledge in Pretrained Language Models. arXiv preprint arXiv:2210.03329 (2022)."},{"key":"e_1_3_2_2_7_1","unstructured":"Zhangyin Feng Weitao Ma Weijiang Yu Lei Huang Haotian Wang Qianglong Chen Weihua Peng Xiaocheng Feng Bing Qin et al. 2023. Trends in Integration of Knowledge and Large Language Models: A Survey and Taxonomy of Methods Benchmarks and Applications. arXiv preprint arXiv:2311.05876 (2023)."},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.3"},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.446"},{"key":"e_1_3_2_2_10_1","volume-title":"Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptors. In NeurIPS 2022 Workshop on Robustness in Sequence Modeling.","author":"Hartvigsen Thomas","year":"2022","unstructured":"Thomas Hartvigsen, Swami Sankaranarayanan, Hamid Palangi, Yoon Kim, and Marzyeh Ghassemi. 2022. Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptors. In NeurIPS 2022 Workshop on Robustness in Sequence Modeling."},{"key":"e_1_3_2_2_11_1","volume-title":"A survey of large language models for healthcare: from data, technology, and applications to accountability and ethics. arXiv preprint arXiv:2310.05694","author":"He Kai","year":"2023","unstructured":"Kai He, Rui Mao, Qika Lin, Yucheng Ruan, Xiang Lan, Mengling Feng, and Erik Cambria. 2023. A survey of large language models for healthcare: from data, technology, and applications to accountability and ethics. arXiv preprint arXiv:2310.05694 (2023)."},{"key":"e_1_3_2_2_12_1","volume-title":"LoRA: Low-Rank Adaptation of Large Language Models. arXiv preprint arXiv:2106.09685","author":"Hu Edward J","year":"2021","unstructured":"Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2021. LoRA: Low-Rank Adaptation of Large Language Models. arXiv preprint arXiv:2106.09685 (2021)."},{"key":"e_1_3_2_2_13_1","volume-title":"Transformer-Patcher: One Mistake Worth One Neuron. In The Eleventh International Conference on Learning Representations.","author":"Huang Zeyu","year":"2023","unstructured":"Zeyu Huang, Yikang Shen, Xiaofeng Zhang, Jie Zhou, Wenge Rong, and Zhang Xiong. 2023. Transformer-Patcher: One Mistake Worth One Neuron. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_2_14_1","unstructured":"Vassilis N. Ioannidis Xiang Song Saurav Manchanda Mufei Li Xiaoqin Pan Da Zheng Xia Ning Xiangxiang Zeng and George Karypis. 2020. DRKG - Drug Repurposing Knowledge Graph for Covid-19."},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.7759\/cureus.39305"},{"key":"e_1_3_2_2_16_1","volume-title":"Camille Elepa no, Maria Madriaga, Rimel Aggabao, Giezel Diaz-Candido, James Maningo, et al.","author":"Kung Tiffany H","year":"2023","unstructured":"Tiffany H Kung, Morgan Cheatham, Arielle Medenilla, Czarina Sillos, Lorie De Leon, Camille Elepa no, Maria Madriaga, Rimel Aggabao, Giezel Diaz-Candido, James Maningo, et al. 2023. Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. PLoS digital health, Vol. 2, 2 (2023), e0000198."},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btz682"},{"key":"e_1_3_2_2_18_1","volume-title":"Contextual Distillation Model for Diversified Recommendation. arXiv preprint arXiv:2406.09021","author":"Li Fan","year":"2024","unstructured":"Fan Li, Xu Si, Shisong Tang, Dingmin Wang, Kunyan Han, Bing Han, Guorui Zhou, Yang Song, and Hechang Chen. 2024. Contextual Distillation Model for Diversified Recommendation. arXiv preprint arXiv:2406.09021 (2024)."},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.397"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3626772.3657722"},{"key":"e_1_3_2_2_21_1","volume-title":"Large language model distilling medication recommendation model. arXiv preprint arXiv:2402.02803","author":"Liu Qidong","year":"2024","unstructured":"Qidong Liu, Xian Wu, Xiangyu Zhao, Yuanshao Zhu, Zijian Zhang, Feng Tian, and Yefeng Zheng. 2024. Large language model distilling medication recommendation model. arXiv preprint arXiv:2402.02803 (2024)."},{"key":"e_1_3_2_2_22_1","volume-title":"Interpreting Key Mechanisms of Factual Recall in Transformer-Based Language Models. arXiv preprint arXiv:2403.19521","author":"Lv Ang","year":"2024","unstructured":"Ang Lv, Kaiyi Zhang, Yuhan Chen, Yulong Wang, Lifeng Liu, Ji-Rong Wen, Jian Xie, and Rui Yan. 2024. Interpreting Key Mechanisms of Factual Recall in Transformer-Based Language Models. arXiv preprint arXiv:2403.19521 (2024)."},{"key":"e_1_3_2_2_23_1","volume-title":"Faithful chain-of-thought reasoning. arXiv preprint arXiv:2301.13379","author":"Lyu Qing","year":"2023","unstructured":"Qing Lyu, Shreya Havaldar, Adam Stein, Li Zhang, Delip Rao, Eric Wong, Marianna Apidianaki, and Chris Callison-Burch. 2023. Faithful chain-of-thought reasoning. arXiv preprint arXiv:2301.13379 (2023)."},{"key":"e_1_3_2_2_24_1","first-page":"17359","article-title":"Locating and Editing Factual Associations in GPT","volume":"35","author":"Meng Kevin","year":"2022","unstructured":"Kevin Meng, David Bau, Alex Andonian, and Yonatan Belinkov. 2022. Locating and Editing Factual Associations in GPT. Advances in Neural Information Processing Systems, Vol. 35 (2022), 17359--17372.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_25_1","volume-title":"Alex Andonian, Yonatan Belinkov, and David Bau.","author":"Meng Kevin","year":"2022","unstructured":"Kevin Meng, Arnab Sen Sharma, Alex Andonian, Yonatan Belinkov, and David Bau. 2022. Mass-Editing Memory in a Transformer. arXiv preprint arXiv:2210.07229 (2022)."},{"key":"e_1_3_2_2_26_1","volume-title":"International Conference on Learning Representations.","author":"Mitchell Eric","year":"2022","unstructured":"Eric Mitchell, Charles Lin, Antoine Bosselut, Chelsea Finn, and Christopher D Manning. 2022. Fast Model Editing at Scale. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_27_1","volume-title":"International Conference on Machine Learning. PMLR, 15817--15831","author":"Mitchell Eric","year":"2022","unstructured":"Eric Mitchell, Charles Lin, Antoine Bosselut, Christopher D Manning, and Chelsea Finn. 2022. Memory-based model editing at scale. In International Conference on Machine Learning. PMLR, 15817--15831."},{"key":"e_1_3_2_2_28_1","volume-title":"Proceedings of the Conference on Health, Inference, and Learning (Proceedings of Machine Learning Research","volume":"260","author":"Pal Ankit","year":"2022","unstructured":"Ankit Pal, Logesh Kumar Umapathi, and Malaikannan Sankarasubbu. 2022. MedMCQA: A Large-scale Multi-Subject Multi-Choice Dataset for Medical domain Question Answering. In Proceedings of the Conference on Health, Inference, and Learning (Proceedings of Machine Learning Research, Vol. 174), Gerardo Flores, George H Chen, Tom Pollard, Joyce C Ho, and Tristan Naumann (Eds.). PMLR, 248--260."},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"crossref","unstructured":"Judea Pearl. 2022. Direct and indirect effects. In Probabilistic and causal inference: the works of Judea Pearl. 373--392.","DOI":"10.1145\/3501714.3501736"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1250"},{"key":"e_1_3_2_2_31_1","volume-title":"International Conference on Learning Representations.","author":"Sun Zhiqing","year":"2018","unstructured":"Zhiqing Sun, Zhi-Hong Deng, Jian-Yun Nie, and Jian Tang. 2018. RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3485447.3511934"},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599797"},{"key":"e_1_3_2_2_34_1","volume-title":"Kabilan Elangovan, Laura Gutierrez, Ting Fang Tan, and Daniel Shu Wei Ting.","author":"Thirunavukarasu Arun James","year":"2023","unstructured":"Arun James Thirunavukarasu, Darren Shu Jeng Ting, Kabilan Elangovan, Laura Gutierrez, Ting Fang Tan, and Daniel Shu Wei Ting. 2023. Large language models in medicine. Nature medicine, Vol. 29, 8 (2023), 1930--1940."},{"key":"e_1_3_2_2_35_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale et al. 2023. Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv preprint arXiv:2307.09288 (2023)."},{"key":"e_1_3_2_2_36_1","volume-title":"Lin (Eds.)","volume":"33","author":"Vig Jesse","year":"2020","unstructured":"Jesse Vig, Sebastian Gehrmann, Yonatan Belinkov, Sharon Qian, Daniel Nevo, Yaron Singer, and Stuart Shieber. 2020. Investigating Gender Bias in Language Models Using Causal Mediation Analysis. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 12388--12401."},{"key":"e_1_3_2_2_37_1","unstructured":"Peng Wang Ningyu Zhang Xin Xie Yunzhi Yao Bozhong Tian Mengru Wang Zekun Xi Siyuan Cheng Kangwei Liu Guozhou Zheng et al. 2023. EasyEdit: An Easy-to-use Knowledge Editing Framework for Large Language Models. arXiv preprint arXiv:2308.07269 (2023)."},{"key":"e_1_3_2_2_38_1","unstructured":"Song Wang Yaochen Zhu Haochen Liu Zaiyi Zheng Chen Chen and Jundong Li. 2023. Knowledge Editing for Large Language Models: A Survey. (2023). arxiv: 2310.16218 [cs.CL]"},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599810"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.174"},{"key":"e_1_3_2_2_41_1","volume-title":"Large language models for generative information extraction: A survey. arXiv preprint arXiv:2312.17617","author":"Xu Derong","year":"2023","unstructured":"Derong Xu, Wei Chen, Wenjun Peng, Chao Zhang, Tong Xu, Xiangyu Zhao, Xian Wu, Yefeng Zheng, and Enhong Chen. 2023. Large language models for generative information extraction: A survey. arXiv preprint arXiv:2312.17617 (2023)."},{"key":"e_1_3_2_2_42_1","volume-title":"Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024","author":"Xu Derong","year":"2024","unstructured":"Derong Xu, Ziheng Zhang, Zhenxi Lin, Xian Wu, Zhihong Zhu, Tong Xu, Xiangyu Zhao, Yefeng Zheng, and Enhong Chen. 2024. Multi-perspective Improvement of Knowledge Graph Completion with Large Language Models. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024). 11956--11968."},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDE55515.2023.10231041"},{"key":"e_1_3_2_2_44_1","volume-title":"Editing Large Language Models: Problems, Methods, and Opportunities. arXiv preprint arXiv:2305.13172","author":"Yao Yunzhi","year":"2023","unstructured":"Yunzhi Yao, Peng Wang, Bozhong Tian, Siyuan Cheng, Zhoubo Li, Shumin Deng, Huajun Chen, and Ningyu Zhang. 2023. Editing Large Language Models: Problems, Methods, and Opportunities. arXiv preprint arXiv:2305.13172 (2023)."},{"key":"e_1_3_2_2_45_1","volume-title":"Cognitive mirage: A review of hallucinations in large language models. arXiv preprint arXiv:2309.06794","author":"Ye Hongbin","year":"2023","unstructured":"Hongbin Ye, Tong Liu, Aijia Zhang, Wei Hua, and Weiqiang Jia. 2023. Cognitive mirage: A review of hallucinations in large language models. arXiv preprint arXiv:2309.06794 (2023)."},{"key":"e_1_3_2_2_46_1","volume-title":"MELO: Enhancing Model Editing with Neuron-Indexed Dynamic LoRA. arXiv preprint arXiv:2312.11795","author":"Yu Lang","year":"2023","unstructured":"Lang Yu, Qin Chen, Jie Zhou, and Liang He. 2023. MELO: Enhancing Model Editing with Neuron-Indexed Dynamic LoRA. arXiv preprint arXiv:2312.11795 (2023)."},{"key":"e_1_3_2_2_47_1","volume-title":"ChatDoctor: A Medical Chat Model Fine-Tuned on a Large Language Model Meta-AI (LLaMA) Using Medical Domain Knowledge. arXiv preprint arXiv:2303.14070","author":"Yunxiang Li","year":"2023","unstructured":"Li Yunxiang, Li Zihan, Zhang Kai, Dan Ruilong, and Zhang You. 2023. ChatDoctor: A Medical Chat Model Fine-Tuned on a Large Language Model Meta-AI (LLaMA) Using Medical Domain Knowledge. arXiv preprint arXiv:2303.14070 (2023)."},{"key":"e_1_3_2_2_48_1","doi-asserted-by":"crossref","unstructured":"Ningyu Zhang Yunzhi Yao Bozhong Tian Peng Wang Shumin Deng Mengru Wang Zekun Xi Shengyu Mao Jintian Zhang Yuansheng Ni et al. 2024. A Comprehensive Study of Knowledge Editing for Large Language Models. arXiv preprint arXiv:2401.01286 (2024).","DOI":"10.18653\/v1\/2024.findings-emnlp.40"},{"key":"e_1_3_2_2_49_1","unstructured":"Yue Zhang Yafu Li Leyang Cui Deng Cai Lemao Liu Tingchen Fu Xinting Huang Enbo Zhao Yu Zhang Yulong Chen et al. 2023. Siren's song in the AI ocean: a survey on hallucination in large language models. arXiv preprint arXiv:2309.01219 (2023)."},{"key":"e_1_3_2_2_50_1","unstructured":"Wayne Xin Zhao Kun Zhou Junyi Li Tianyi Tang Xiaolei Wang Yupeng Hou Yingqian Min Beichen Zhang Junjie Zhang Zican Dong et al. 2023. A survey of large language models. arXiv preprint arXiv:2303.18223 (2023)."},{"key":"e_1_3_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.296"},{"key":"e_1_3_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539201"},{"key":"e_1_3_2_2_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3485447.3512099"},{"key":"e_1_3_2_2_54_1","unstructured":"Hongjian Zhou Boyang Gu Xinyu Zou Yiru Li Sam S Chen Peilin Zhou Junling Liu Yining Hua Chengfeng Mao Xian Wu et al. 2023. A survey of large language models in medicine: Progress application and challenge. arXiv preprint arXiv:2311.05112 (2023)."}],"event":{"name":"CIKM '24: The 33rd ACM International Conference on Information and Knowledge Management","location":"Boise ID USA","acronym":"CIKM '24","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval"]},"container-title":["Proceedings of the 33rd ACM International Conference on Information and Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3627673.3679673","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3627673.3679673","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:58:13Z","timestamp":1750294693000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3627673.3679673"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,21]]},"references-count":54,"alternative-id":["10.1145\/3627673.3679673","10.1145\/3627673"],"URL":"https:\/\/doi.org\/10.1145\/3627673.3679673","relation":{},"subject":[],"published":{"date-parts":[[2024,10,21]]},"assertion":[{"value":"2024-10-21","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}