{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,3]],"date-time":"2025-12-03T17:24:06Z","timestamp":1764782646815,"version":"3.46.0"},"publisher-location":"New York, NY, USA","reference-count":56,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,12,7]]},"DOI":"10.1145\/3767695.3769672","type":"proceedings-article","created":{"date-parts":[[2025,12,3]],"date-time":"2025-12-03T17:14:58Z","timestamp":1764782098000},"page":"453-458","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Dynamic and Parametric Retrieval Augmented Generation"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8718-9402","authenticated-orcid":false,"given":"Weihang","family":"Su","sequence":"first","affiliation":[{"name":"Department of Computer Science and Technology, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6858-5303","authenticated-orcid":false,"given":"Qian","family":"Dong","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5030-709X","authenticated-orcid":false,"given":"Qingyao","family":"Ai","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0140-4512","authenticated-orcid":false,"given":"Yiqun","family":"Liu","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Tsinghua University, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2025,12,6]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Gianluca Antolini. 2025. Experimental Study on Retrieval-Augmented Generation: Engineering and Evaluation of a Custom RAG system for Open-Domain QA. (2025)."},{"key":"e_1_3_2_1_2_1","volume-title":"International Conference on Learning Representations (ICLR).","author":"Asai Akari","year":"2024","unstructured":"Akari Asai, Zeqiu Wu, Yizhong Wang, Avirup Sil, and Hannaneh Hajishirzi. 2024. Self-Reflective Retrieval-Augmented Generation (Self-RAG). In International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_3_1","volume-title":"International conference on machine learning. PMLR, 2206-2240","author":"Borgeaud Sebastian","year":"2022","unstructured":"Sebastian Borgeaud, Arthur Mensch, Jordan Hoffmann, Trevor Cai, Eliza Rutherford, Katie Millican, George Bm Van Den Driessche, Jean-Baptiste Lespiau, Bogdan Damoc, Aidan Clark, et al., 2022. Improving language models by retrieving from trillions of tokens. In International conference on machine learning. PMLR, 2206-2240."},{"key":"e_1_3_2_1_4_1","unstructured":"Tom Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared D Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et al. 2020. Language models are few-shot learners. Advances in neural information processing systems Vol. 33 (2020) 1877-1901."},{"key":"e_1_3_2_1_5_1","volume-title":"Privacy-Preserving Reasoning with Knowledge-Distilled Parametric Retrieval Augmented Generation. arXiv preprint arXiv:2509.01088","author":"Chen Jinwen","year":"2025","unstructured":"Jinwen Chen, Hainan Zhang, Liang Pang, Yongxin Tong, Haibo Zhou, Yuan Zhan, Wei Lin, and Zhiming Zheng. 2025. Privacy-Preserving Reasoning with Knowledge-Distilled Parametric Retrieval Augmented Generation. arXiv preprint arXiv:2509.01088 (2025)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3488560.3502185"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3696410.3714608"},{"key":"e_1_3_2_1_8_1","volume-title":"Alphaedit: Null-space constrained knowledge editing for language models. arXiv preprint arXiv:2410.02355","author":"Fang Junfeng","year":"2024","unstructured":"Junfeng Fang, Houcheng Jiang, Kun Wang, Yunshan Ma, Shi Jie, Xiang Wang, Xiangnan He, and Tat-Seng Chua. 2024a. Alphaedit: Null-space constrained knowledge editing for language models. arXiv preprint arXiv:2410.02355 (2024)."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3626772.3657743"},{"key":"e_1_3_2_1_10_1","volume-title":"arXiv preprint arXiv:2507.05346","author":"Fleshman William","year":"2025","unstructured":"William Fleshman and Benjamin Van Durme. 2025. LoRA-Augmented Generation (LAG) for Knowledge-Intensive Language Tasks. arXiv preprint arXiv:2507.05346 (2025)."},{"key":"e_1_3_2_1_11_1","volume-title":"Condenser: a pre-training architecture for dense retrieval. arXiv preprint arXiv:2104.08253","author":"Gao Luyu","year":"2021","unstructured":"Luyu Gao and Jamie Callan. 2021. Condenser: a pre-training architecture for dense retrieval. arXiv preprint arXiv:2104.08253 (2021)."},{"key":"e_1_3_2_1_12_1","volume-title":"Don't stop pretraining: Adapt language models to domains and tasks. arXiv preprint arXiv:2004.10964","author":"Gururangan Suchin","year":"2020","unstructured":"Suchin Gururangan, Ana Marasovi\u0107, Swabha Swayamdipta, Kyle Lo, Iz Beltagy, Doug Downey, and Noah A Smith. 2020. Don't stop pretraining: Adapt language models to domains and tasks. arXiv preprint arXiv:2004.10964 (2020)."},{"key":"e_1_3_2_1_13_1","volume-title":"International conference on machine learning. PMLR, 3929-3938","author":"Guu Kelvin","year":"2020","unstructured":"Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, and Mingwei Chang. 2020. Retrieval augmented language model pre-training. In International conference on machine learning. PMLR, 3929-3938."},{"key":"e_1_3_2_1_14_1","volume-title":"LoRA: Low-Rank Adaptation of Large Language Models. In International Conference on Learning Representations.","author":"Hu Edward J","year":"2022","unstructured":"Edward J Hu, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al., 2022. LoRA: Low-Rank Adaptation of Large Language Models. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_15_1","volume-title":"Leveraging passage retrieval with generative models for open domain question answering. arXiv preprint arXiv:2007.01282","author":"Izacard Gautier","year":"2020","unstructured":"Gautier Izacard and Edouard Grave. 2020. Leveraging passage retrieval with generative models for open domain question answering. arXiv preprint arXiv:2007.01282 (2020)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3571730"},{"key":"e_1_3_2_1_17_1","volume-title":"Retrieval as attention: End-to-end learning of retrieval and reading within a single transformer. arXiv preprint arXiv:2212.02027","author":"Jiang Zhengbao","year":"2022","unstructured":"Zhengbao Jiang, Luyu Gao, Jun Araki, Haibo Ding, Zhiruo Wang, Jamie Callan, and Graham Neubig. 2022. Retrieval as attention: End-to-end learning of retrieval and reading within a single transformer. arXiv preprint arXiv:2212.02027 (2022)."},{"key":"e_1_3_2_1_18_1","volume-title":"Active retrieval augmented generation. arXiv preprint arXiv:2305.06983","author":"Jiang Zhengbao","year":"2023","unstructured":"Zhengbao Jiang, Frank F Xu, Luyu Gao, Zhiqing Sun, Qian Liu, Jane Dwivedi-Yu, Yiming Yang, Jamie Callan, and Graham Neubig. 2023. Active retrieval augmented generation. arXiv preprint arXiv:2305.06983 (2023)."},{"key":"e_1_3_2_1_19_1","volume-title":"Dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2004.04906","author":"Karpukhin Vladimir","year":"2020","unstructured":"Vladimir Karpukhin, Barlas O\u011fuz, Sewon Min, Patrick Lewis, Ledell Wu, Sergey Edunov, Danqi Chen, and Wen-tau Yih. 2020. Dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2004.04906 (2020)."},{"key":"e_1_3_2_1_20_1","volume-title":"Same task, more tokens: the impact of input length on the reasoning performance of large language models. arXiv preprint arXiv:2402.14848","author":"Levy Mosh","year":"2024","unstructured":"Mosh Levy, Alon Jacoby, and Yoav Goldberg. 2024. Same task, more tokens: the impact of input length on the reasoning performance of large language models. arXiv preprint arXiv:2402.14848 (2024)."},{"key":"e_1_3_2_1_21_1","first-page":"9459","article-title":"Retrieval-augmented generation for knowledge-intensive nlp tasks","volume":"33","author":"Lewis Patrick","year":"2020","unstructured":"Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich K\u00fcttler, Mike Lewis, Wen-tau Yih, Tim Rockt\u00e4schel, et al., 2020. Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems, Vol. 33 (2020), 9459-9474.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_22_1","volume-title":"Towards better web search performance: pre-training, fine-tuning and learning to rank. arXiv preprint arXiv:2303.04710","author":"Li Haitao","year":"2023","unstructured":"Haitao Li, Jia Chen, Weihang Su, Qingyao Ai, and Yiqun Liu. 2023. Towards better web search performance: pre-training, fine-tuning and learning to rank. arXiv preprint arXiv:2303.04710 (2023)."},{"key":"e_1_3_2_1_23_1","unstructured":"Aixin Liu Bei Feng Bing Xue Bingxuan Wang Bochao Wu Chengda Lu Chenggang Zhao Chengqi Deng Chenyu Zhang Chong Ruan et al. 2024a. Deepseek-v3 technical report. arXiv preprint arXiv:2412.19437 (2024)."},{"key":"e_1_3_2_1_24_1","volume-title":"Cong Zhang, and Yong Liu.","author":"Liu Huanshuo","year":"2024","unstructured":"Huanshuo Liu, Hao Zhang, Zhijiang Guo, Kuicai Dong, Xiangyang Li, Yi Quan Lee, Cong Zhang, and Yong Liu. 2024c. CtrlA: Adaptive Retrieval-Augmented Generation via Probe-Guided Control. arXiv preprint arXiv:2405.18727 (2024)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00638"},{"key":"e_1_3_2_1_26_1","volume-title":"DRAE: Dynamic Retrieval-Augmented Expert Networks for Lifelong Learning and Task Adaptation in Robotics. arXiv preprint arXiv:2507.04661","author":"Long Yayu","year":"2025","unstructured":"Yayu Long, Kewei Chen, Long Jin, and Mingsheng Shang. 2025. DRAE: Dynamic Retrieval-Augmented Expert Networks for Lifelong Learning and Task Adaptation in Robotics. arXiv preprint arXiv:2507.04661 (2025)."},{"key":"e_1_3_2_1_27_1","volume-title":"CaseEncoder: A Knowledge-enhanced Pre-trained Model for Legal Case Encoding. arXiv preprint arXiv:2305.05393","author":"Ma Yixiao","year":"2023","unstructured":"Yixiao Ma, Yueyue Wu, Weihang Su, Qingyao Ai, and Yiqun Liu. 2023. CaseEncoder: A Knowledge-enhanced Pre-trained Model for Legal Case Encoding. arXiv preprint arXiv:2305.05393 (2023)."},{"key":"e_1_3_2_1_28_1","unstructured":"Neel Nanda Senthooran Rajamanoharan J\u00e1nos Kram\u00e1r and Rohin Shah. 2023. Fact Finding: Attempting to Reverse-Engineer Factual Recall on the Neuron Level. https:\/\/www.lesswrong.com\/posts\/iGuwZTHWb6DFY3sKB\/fact-finding-attempting-to-reverse-engineer-factual-recall Accessed: 2025-01-24."},{"key":"e_1_3_2_1_29_1","volume-title":"A survey of hallucination in large foundation models. arXiv preprint arXiv:2309.05922","author":"Rawte Vipula","year":"2023","unstructured":"Vipula Rawte, Amit Sheth, and Amitava Das. 2023. A survey of hallucination in large foundation models. arXiv preprint arXiv:2309.05922 (2023)."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1561\/1500000019"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3626772.3657733"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i17.29869"},{"key":"e_1_3_2_1_33_1","volume-title":"Caseformer: Pre-training for Legal Case Retrieval. arXiv preprint arXiv:2311.00333","author":"Su Weihang","year":"2023","unstructured":"Weihang Su, Qingyao Ai, Yueyue Wu, Yixiao Ma, Haitao Li, and Yiqun Liu. 2023a. Caseformer: Pre-training for Legal Case Retrieval. arXiv preprint arXiv:2311.00333 (2023)."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-emnlp.625"},{"key":"e_1_3_2_1_35_1","volume-title":"Thuir2 at ntcir-16 session search (ss) task. arXiv preprint arXiv:2307.00250","author":"Su Weihang","year":"2023","unstructured":"Weihang Su, Xiangsheng Li, Yiqun Liu, Min Zhang, and Shaoping Ma. 2023b. Thuir2 at ntcir-16 session search (ss) task. arXiv preprint arXiv:2307.00250 (2023)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3673791.3698403"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","unstructured":"Weihang Su Yichen Tang Qingyao Ai Zhijing Wu and Yiqun Liu. 2024d. DRAGIN: Dynamic Retrieval Augmented Generation based on the Real-time Information Needs of Large Language Models. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) Lun-Wei Ku Andre Martins and Vivek Srikumar (Eds.). Association for Computational Linguistics Bangkok Thailand 12991-13013. https:\/\/doi.org\/10.18653\/v1\/2024.acl-long.702","DOI":"10.18653\/v1\/2024.acl-long.702"},{"key":"e_1_3_2_1_38_1","volume-title":"Parametric Retrieval-Augmented Generation. arXiv preprint arXiv:2501.15915","author":"Su Weihang","year":"2025","unstructured":"Weihang Su, Yichen Tang, Qingyao Ai, Junxi Yan, Changyue Wang, Hongning Wang, Ziyi Ye, Yujia Zhou, and Yiqun Liu. 2025a. Parametric Retrieval-Augmented Generation. arXiv preprint arXiv:2501.15915 (2025)."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3726302.3729957"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-acl.854"},{"key":"e_1_3_2_1_41_1","unstructured":"Weihang Su Changyue Wang Anzhe Xie Qingyao Ai Yiran Hu and Yiqun Liu. 2024 f. LegalAID: A Large Language Model for the Chinese Legal Field. https:\/\/github.com\/oneal2000\/LegalAID."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3726302.3730295"},{"key":"e_1_3_2_1_43_1","volume-title":"Better wit than wealth: Dynamic Parametric Retrieval Augmented Generation for Test-time Knowledge Enhancement. arXiv preprint arXiv:2503.23895","author":"Tan Yuqiao","year":"2025","unstructured":"Yuqiao Tan, Shizhu He, Huanxuan Liao, Jun Zhao, and Kang Liu. 2025. Better wit than wealth: Dynamic Parametric Retrieval Augmented Generation for Test-time Knowledge Enhancement. arXiv preprint arXiv:2503.23895 (2025)."},{"key":"e_1_3_2_1_44_1","volume-title":"Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, et al., 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_1_45_1","volume-title":"RbFT: Robust Fine-tuning for Retrieval-Augmented Generation against Retrieval Defects. arXiv preprint arXiv:2501.18365","author":"Tu Yiteng","year":"2025","unstructured":"Yiteng Tu, Weihang Su, Yujia Zhou, Yiqun Liu, and Qingyao Ai. 2025. RbFT: Robust Fine-tuning for Retrieval-Augmented Generation against Retrieval Defects. arXiv preprint arXiv:2501.18365 (2025)."},{"key":"e_1_3_2_1_46_1","volume-title":"Knowledge Editing through Chain-of-Thought. arXiv preprint arXiv:2412.17727","author":"Wang Changyue","year":"2024","unstructured":"Changyue Wang, Weihang Su, Qingyao Ai, and Yiqun Liu. 2024a. Knowledge Editing through Chain-of-Thought. arXiv preprint arXiv:2412.17727 (2024)."},{"key":"e_1_3_2_1_47_1","volume-title":"Decoupling Reasoning and Knowledge Injection for In-Context Knowledge Editing. arXiv preprint arXiv:2506.00536","author":"Wang Changyue","year":"2025","unstructured":"Changyue Wang, Weihang Su, Qingyao Ai, and Yiqun Liu. 2025a. Decoupling Reasoning and Knowledge Injection for In-Context Knowledge Editing. arXiv preprint arXiv:2506.00536 (2025)."},{"key":"e_1_3_2_1_48_1","volume-title":"Joint Evaluation of Answer and Reasoning Consistency for Hallucination Detection in Large Reasoning Models. arXiv preprint arXiv:2506.04832","author":"Wang Changyue","year":"2025","unstructured":"Changyue Wang, Weihang Su, Qingyao Ai, and Yiqun Liu. 2025b. Joint Evaluation of Answer and Reasoning Consistency for Hallucination Detection in Large Reasoning Models. arXiv preprint arXiv:2506.04832 (2025)."},{"key":"e_1_3_2_1_49_1","volume-title":"LeKUBE: A Legal Knowledge Update BEnchmark. arXiv preprint arXiv:2407.14192","author":"Wang Changyue","year":"2024","unstructured":"Changyue Wang, Weihang Su, Hu Yiran, Qingyao Ai, Yueyue Wu, Cheng Luo, Yiqun Liu, Min Zhang, and Shaoping Ma. 2024b. LeKUBE: A Legal Knowledge Update BEnchmark. arXiv preprint arXiv:2407.14192 (2024)."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3698590","article-title":"Knowledge editing for large language models: A survey","volume":"57","author":"Wang Song","year":"2024","unstructured":"Song Wang, Yaochen Zhu, Haochen Liu, Zaiyi Zheng, Chen Chen, and Jundong Li. 2024c. Knowledge editing for large language models: A survey. Comput. Surveys, Vol. 57, 3 (2024), 1-37.","journal-title":"Comput. Surveys"},{"key":"e_1_3_2_1_51_1","unstructured":"An Yang Baosong Yang Beichen Zhang Binyuan Hui Bo Zheng Bowen Yu Chengyuan Li Dayiheng Liu Fei Huang Haoran Wei et al. 2024. Qwen2. 5 Technical Report. arXiv preprint arXiv:2412.15115 (2024)."},{"key":"e_1_3_2_1_52_1","volume-title":"Seakr: Self-aware knowledge retrieval for adaptive retrieval augmented generation. arXiv preprint arXiv:2406.19215","author":"Yao Zijun","year":"2024","unstructured":"Zijun Yao, Weijian Qi, Liangming Pan, Shulin Cao, Linmei Hu, Weichuan Liu, Lei Hou, and Juanzi Li. 2024. Seakr: Self-aware knowledge retrieval for adaptive retrieval augmented generation. arXiv preprint arXiv:2406.19215 (2024)."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3637874"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.191"},{"key":"e_1_3_2_1_55_1","volume-title":"Statistical language models for information retrieval. Synthesis lectures on human language technologies","author":"Zhai ChengXiang","year":"2008","unstructured":"ChengXiang Zhai. 2008. Statistical language models for information retrieval. Synthesis lectures on human language technologies, Vol. 1, 1 (2008), 1-141."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462880"}],"event":{"name":"SIGIR-AP 2025:Annual International ACM SIGIR Conference on Research and Development in Information Retrieval in the Asia Pacific Region","location":"Xi'an China","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval"]},"container-title":["Proceedings of the 2025 Annual International ACM SIGIR Conference on Research and Development in Information Retrieval in the Asia Pacific Region"],"original-title":[],"deposited":{"date-parts":[[2025,12,3]],"date-time":"2025-12-03T17:18:35Z","timestamp":1764782315000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3767695.3769672"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,6]]},"references-count":56,"alternative-id":["10.1145\/3767695.3769672","10.1145\/3767695"],"URL":"https:\/\/doi.org\/10.1145\/3767695.3769672","relation":{},"subject":[],"published":{"date-parts":[[2025,12,6]]},"assertion":[{"value":"2025-12-06","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}