{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T17:54:50Z","timestamp":1775325290024,"version":"3.50.1"},"reference-count":179,"publisher":"Association for Computing Machinery (ACM)","issue":"3","license":[{"start":{"date-parts":[[2024,11,11]],"date-time":"2024-11-11T00:00:00Z","timestamp":1731283200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"National Science Foundation","award":["IIS-2006844, IIS-2144209, IIS-2223769, CNS2154962, and BCS-2228534"],"award-info":[{"award-number":["IIS-2006844, IIS-2144209, IIS-2223769, CNS2154962, and BCS-2228534"]}]},{"name":"Commonwealth Cyber Initiative awards","award":["VV-1Q23-007, HV-2Q23-003, and VV-1Q24-011"],"award-info":[{"award-number":["VV-1Q23-007, HV-2Q23-003, and VV-1Q24-011"]}]},{"name":"JP Morgan Chase Faculty Research Award, the Cisco Faculty Research Award, the Jefferson Lab subcontract, and the UVA 4-VA collaborative research grant"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Comput. Surv."],"published-print":{"date-parts":[[2025,3,31]]},"abstract":"<jats:p>Large Language Models (LLMs) have recently transformed both the academic and industrial landscapes due to their remarkable capacity to understand, analyze, and generate texts based on their vast knowledge and reasoning ability. Nevertheless, one major drawback of LLMs is their substantial computational cost for pre-training due to their unprecedented amounts of parameters. The disadvantage is exacerbated when new knowledge frequently needs to be introduced into the pre-trained model. Therefore, it is imperative to develop effective and efficient techniques to update pre-trained LLMs. Traditional methods encode new knowledge in pre-trained LLMs through direct fine-tuning. However, naively re-training LLMs can be computationally intensive and risks degenerating valuable pre-trained knowledge irrelevant to the update in the model. Recently,<jats:bold>Knowledge-based Model Editing<\/jats:bold>(KME), also known as<jats:bold>Knowledge Editing<\/jats:bold>or<jats:bold>Model Editing<\/jats:bold>, has attracted increasing attention, which aims at precisely modifying the LLMs to incorporate specific knowledge, without negatively influencing other irrelevant knowledge. In this survey, we aim at providing a comprehensive and in-depth overview of recent advances in the field of KME. We first introduce a general formulation of KME to encompass different KME strategies. Afterward, we provide an innovative taxonomy of KME techniques based on how the new knowledge is introduced into pre-trained LLMs, and investigate existing KME strategies while analyzing key insights, advantages, and limitations of methods from each category. Moreover, representative metrics, datasets, and applications of KME are introduced accordingly. Finally, we provide an in-depth analysis regarding the practicality and remaining challenges of KME and suggest promising research directions for further advancement in this field.<\/jats:p>","DOI":"10.1145\/3698590","type":"journal-article","created":{"date-parts":[[2024,10,7]],"date-time":"2024-10-07T10:18:29Z","timestamp":1728296309000},"page":"1-37","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":41,"title":["Knowledge Editing for Large Language Models: A Survey"],"prefix":"10.1145","volume":"57","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1273-7694","authenticated-orcid":false,"given":"Song","family":"Wang","sequence":"first","affiliation":[{"name":"University of Virginia, Charlottesville, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6266-2788","authenticated-orcid":false,"given":"Yaochen","family":"Zhu","sequence":"additional","affiliation":[{"name":"University of Virginia, Charlottesville, United States"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-7515-2166","authenticated-orcid":false,"given":"Haochen","family":"Liu","sequence":"additional","affiliation":[{"name":"University of Virginia, Charlottesville, United States"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-0685-0057","authenticated-orcid":false,"given":"Zaiyi","family":"Zheng","sequence":"additional","affiliation":[{"name":"University of Virginia, Charlottesville, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7099-7905","authenticated-orcid":false,"given":"Chen","family":"Chen","sequence":"additional","affiliation":[{"name":"University of Virginia, Charlottesville, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1878-817X","authenticated-orcid":false,"given":"Jundong","family":"Li","sequence":"additional","affiliation":[{"name":"University of Virginia, Charlottesville, United States"}]}],"member":"320","published-online":{"date-parts":[[2024,11,11]]},"reference":[{"key":"e_1_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.1145\/3461702.3462624"},{"key":"e_1_3_2_3_2","doi-asserted-by":"crossref","unstructured":"Armen Aghajanyan Sonal Gupta and Luke Zettlemoyer. 2021. Intrinsic dimensionality explains the effectiveness of language model fine-tuning. In Proceedings of the Annual Meeting of the Association for Computational Linguistics.","DOI":"10.18653\/v1\/2021.acl-long.568"},{"key":"e_1_3_2_4_2","doi-asserted-by":"crossref","unstructured":"James A. Anderson. 1972. A simple neural network generating an interactive memory. Mathematical Biosciences 14 (1972) 197\u2013220.","DOI":"10.1016\/0025-5564(72)90075-2"},{"key":"e_1_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.279"},{"key":"e_1_3_2_6_2","unstructured":"Robert L. Logan IV au2 Alexandre Passos Sameer Singh and Ming-Wei Chang. 2022. FRUIT: Faithfully reflecting updated information in text. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics."},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1186\/s13054-023-04393-x"},{"key":"e_1_3_2_8_2","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","year":"2022","unstructured":"Michiel Bakker, Martin Chadwick, Hannah Sheahan, Michael Tessler, Lucy Campbell-Gillingham, Jan Balaguer, Nat McAleese, Amelia Glaese, John Aslanides, Matt Botvinick, and Christopher Summerfield. 2022. Fine-tuning language models to find agreement among humans with diverse preferences. In Proceedings of the Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_21"},{"key":"e_1_3_2_10_2","unstructured":"Nora Belrose David Schneider-Joseph Shauli Ravfogel Ryan Cotterell Edward Raff and Stella Biderman. 2023. LEACE: Perfect linear concept erasure in closed form. In International Conference on Learning Representations."},{"key":"e_1_3_2_11_2","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","year":"2020","unstructured":"Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language models are few-shot learners. In Proceedings of the Advances in Neural Information Processing Systems. 1877\u20131901."},{"key":"e_1_3_2_12_2","doi-asserted-by":"crossref","unstructured":"Ilias Chalkidis Tommaso Pasini Sheng Zhang Letizia Tomada Sebastian Schwemer and Anders S\u00f8gaard. 2022. FairLex: A multilingual benchmark for evaluating fairness in legal text processing. In Proceedings of the Annual Meeting of the Association for Computational Linguistics.","DOI":"10.18653\/v1\/2022.acl-long.301"},{"key":"e_1_3_2_13_2","doi-asserted-by":"crossref","unstructured":"Yupeng Chang Xu Wang Jindong Wang Yuan Wu Linyi Yang Kaijie Zhu Hao Chen Xiaoyuan Yi Cunxiang Wang Yidong Wang Wei Ye Yue Zhang Yi Chang Philip S. Yu Qiang Yang and Xing Xie. 2024. A survey on evaluation of large language models. ACM Transactions on Intelligent Systems and Technology 15 3 (2024) 1\u201345.","DOI":"10.1145\/3641289"},{"key":"e_1_3_2_14_2","doi-asserted-by":"crossref","unstructured":"Sanyuan Chen Yutai Hou Yiming Cui Wanxiang Che Ting Liu and Xiangzhan Yu. 2020. Recall and learn: Fine-tuning deep pretrained language models with less forgetting. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2020.emnlp-main.634"},{"key":"e_1_3_2_15_2","unstructured":"Xinlei Chen Hao Fang Tsung-Yi Lin Ramakrishna Vedantam Saurabh Gupta Piotr Doll\u00e1r and C. Lawrence Zitnick. 2015. Microsoft coco captions: Data collection and evaluation server. arXiv preprint arXiv:1504.00325 (2015)."},{"key":"e_1_3_2_16_2","unstructured":"Yingfa Chen Zhengyan Zhang Xu Han Chaojun Xiao Zhiyuan Liu Chen Chen Kuai Li Tao Yang and Maosong Sun. 2024. Robust and scalable model editing for large language models. In The International Conference on Computational Linguistics."},{"key":"e_1_3_2_17_2","doi-asserted-by":"crossref","unstructured":"Siyuan Cheng Bozhong Tian Qingbin Liu Xi Chen Yongheng Wang Huajun Chen and Ningyu Zhang. 2023. Can we edit multimodal large language models? In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2023.emnlp-main.856"},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i16.29737"},{"key":"e_1_3_2_19_2","doi-asserted-by":"crossref","unstructured":"Cheng-Han Chiang and Hung-yi Lee. 2023. Can large language models be an alternative to human evaluations?arXiv preprint arXiv:2305.01937 (2023).","DOI":"10.18653\/v1\/2023.acl-long.870"},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.1186\/s40537-022-00561-y"},{"key":"e_1_3_2_21_2","unstructured":"Hyung Won Chung Le Hou Shayne Longpre Barret Zoph Yi Tay William Fedus Yunxuan Li Xuezhi Wang Mostafa Dehghani Siddhartha Brahma Albert Webson Shixiang Shane Gu Zhuyun Dai Mirac Suzgun Xinyun Chen Aakanksha Chowdhery Alex Castro-Ros Marie Pellat Kevin Robinson Dasha Valter Sharan Narang Gaurav Mishra Adams Yu Vincent Zhao Yanping Huang Andrew Dai Hongkun Yu Slav Petrov Ed H. Chi Jeff Dean Jacob Devlin Adam Roberts Denny Zhou Quoc V. Le and Jason Wei. 2024. Scaling instruction-finetuned language models. Journal of Machine Learning Research 25 70 (2024) 1\u201353."},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00644"},{"key":"e_1_3_2_23_2","doi-asserted-by":"crossref","unstructured":"Damai Dai Li Dong Yaru Hao Zhifang Sui Baobao Chang and Furu Wei. 2022. Knowledge neurons in pretrained transformers. In Proceedings of the Annual Meeting of the Association for Computational Linguistics.","DOI":"10.18653\/v1\/2022.acl-long.581"},{"key":"e_1_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-44696-2_60"},{"key":"e_1_3_2_25_2","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287572"},{"key":"e_1_3_2_26_2","doi-asserted-by":"crossref","unstructured":"Nicola De Cao Wilker Aziz and Ivan Titov. 2021. Editing factual knowledge in language models. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2021.emnlp-main.522"},{"key":"e_1_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_3_2_28_2","unstructured":"Jacob Devlin Ming-Wei Chang Kenton Lee and Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics."},{"key":"e_1_3_2_29_2","article-title":"Unified language model pre-training for natural language understanding and generation","author":"Dong Li","year":"2019","unstructured":"Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, and Hsiao-Wuen Hon. 2019. Unified language model pre-training for natural language understanding and generation. Advances in Neural Information Processing Systems 32 (2019).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_30_2","doi-asserted-by":"crossref","unstructured":"Qingxiu Dong Damai Dai Yifan Song Jingjing Xu Zhifang Sui and Lei Li. 2022. Calibrating factual knowledge in pretrained language models. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2022.findings-emnlp.438"},{"key":"e_1_3_2_31_2","unstructured":"Qingxiu Dong Lei Li Damai Dai Ce Zheng Zhiyong Wu Baobao Chang Xu Sun Jingjing Xu and Zhifang Sui. 2022. A survey for in-context learning. arXiv preprint arXiv:2301.00234 (2022)."},{"key":"e_1_3_2_32_2","unstructured":"Yann Dubois Chen Xuechen Li Rohan Taori Tianyi Zhang Ishaan Gulrajani Jimmy Ba Carlos Guestrin Percy S. Liang and Tatsunori B. Hashimoto. 2024. Alpacafarm: A simulation framework for methods that learn from human feedback. In Proceedings of the Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_33_2","article-title":"Measuring and improving consistency in pretrained language models","author":"Elazar Yanai","year":"2021","unstructured":"Yanai Elazar, Nora Kassner, Shauli Ravfogel, Abhilasha Ravichander, Eduard Hovy, Hinrich Sch\u00fctze, and Yoav Goldberg. 2021. Measuring and improving consistency in pretrained language models. Transactions of the Association for Computational Linguistics 9 (2021), 1012\u20131031.","journal-title":"Transactions of the Association for Computational Linguistics"},{"key":"e_1_3_2_34_2","volume-title":"Proceedings of the Eleventh International Conference on Language Resources and Evaluation","author":"Elsahar Hady","year":"2018","unstructured":"Hady Elsahar, Pavlos Vougiouklis, Arslen Remaci, Christophe Gravier, Jonathon Hare, Frederique Laforest, and Elena Simperl. 2018. T-rex: A large scale alignment of natural language with knowledge base triples. In Proceedings of the Eleventh International Conference on Language Resources and Evaluation."},{"key":"e_1_3_2_35_2","unstructured":"Wenqi Fan Zihuai Zhao Jiatong Li Yunqing Liu Xiaowei Mei Yiqi Wang Jiliang Tang and Qing Li. 2023. Recommender systems in the era of large language models (llms). arXiv preprint arXiv:2307.02046 (2023)."},{"key":"e_1_3_2_36_2","doi-asserted-by":"publisher","DOI":"10.1093\/bib\/bbaa110"},{"key":"e_1_3_2_37_2","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Finn Chelsea","year":"2017","unstructured":"Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the International Conference on Machine Learning."},{"key":"e_1_3_2_38_2","doi-asserted-by":"crossref","unstructured":"Boris A. Galitsky. 2023. Truth-O-Meter: Collaborating with LLM in fighting its hallucinations. Preprints (2023).","DOI":"10.20944\/preprints202307.1723.v1"},{"key":"e_1_3_2_39_2","doi-asserted-by":"crossref","unstructured":"Govind Gangadhar and Karl Stratos. 2024. Model editing by pure fine-tuning. arXiv:2402.11078. Retrieved from https:\/\/arxiv.org\/abs\/2402.11078","DOI":"10.18653\/v1\/2024.findings-acl.352"},{"key":"e_1_3_2_40_2","unstructured":"Deep Ganguli Liane Lovitt Jackson Kernion Amanda Askell Yuntao Bai Saurav Kadavath Ben Mann Ethan Perez Nicholas Schiefer Kamal Ndousse Andy Jones Sam Bowman Anna Chen Tom Conerly Nova DasSarma Dawn Drain Nelson Elhage Sheer El-Showk Stanislav Fort Zac Hatfield-Dodds Tom Henighan Danny Hernandez Tristan Hume Josh Jacobson Scott Johnston Shauna Kravec Catherine Olsson Sam Ringer Eli Tran-Johnson Dario Amodei Tom Brown Nicholas Joseph Sam McCandlish Chris Olah Jared Kaplan and Jack Clark. 2022. Red teaming language models to reduce harms: Methods scaling behaviors and lessons learned. arXiv preprint arXiv:2209.07858 (2022)."},{"key":"e_1_3_2_41_2","unstructured":"Yunfan Gao Yun Xiong Xinyu Gao Kangxiang Jia Jinliu Pan Yuxi Bi Yi Dai Jiawei Sun and Haofen Wang. 2023. Retrieval-augmented generation for large language models: A survey. arXiv preprint arXiv:2312.10997 (2023)."},{"key":"e_1_3_2_42_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7952261"},{"key":"e_1_3_2_43_2","doi-asserted-by":"crossref","unstructured":"Mor Geva Avi Caciularu Kevin Wang and Yoav Goldberg. 2022. Transformer feed-forward layers build predictions by promoting concepts in the vocabulary space. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2022.emnlp-main.3"},{"key":"e_1_3_2_44_2","doi-asserted-by":"crossref","unstructured":"Mor Geva Roei Schuster Jonathan Berant and Omer Levy. 2021. Transformer feed-forward layers are key-value memories. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2021.emnlp-main.446"},{"key":"e_1_3_2_45_2","unstructured":"Amelia Glaese Nat McAleese Maja Tr\u0229bacz John Aslanides Vlad Firoiu Timo Ewalds Maribeth Rauh Laura Weidinger Martin Chadwick Phoebe Thacker Lucy Campbell-Gillingham Jonathan Uesato Po-Sen Huang Ramona Comanescu Fan Yang Abigail See Sumanth Dathathri Rory Greig Charlie Chen Doug Fritz Jaume Sanchez Elias Richard Green So\u0148a Mokr\u00e1 Nicholas Fernando Boxi Wu Rachel Foley Susannah Young Iason Gabriel William Isaac John Mellor Demis Hassabis Koray Kavukcuoglu Lisa Anne Hendricks and Geoffrey Irving. 2022. Improving alignment of dialogue agents via targeted human judgements. arXiv preprint arXiv:2209.14375 (2022)."},{"key":"e_1_3_2_46_2","doi-asserted-by":"crossref","unstructured":"Ian Goodfellow Jean Pouget-Abadie Mehdi Mirza Bing Xu David Warde-Farley Sherjil Ozair Aaron Courville and Yoshua Bengio. 2020. Generative adversarial networks. Commun. ACM 63 11 (2020) 139\u2013144.","DOI":"10.1145\/3422622"},{"key":"e_1_3_2_47_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.670"},{"key":"e_1_3_2_48_2","unstructured":"Hengrui Gu Kaixiong Zhou Xiaotian Han Ninghao Liu Ruobing Wang and Xin Wang. 2023. Pokemqa: Programmable knowledge editing for multi-hop question answering. arXiv preprint arXiv:2312.15194 (2023)."},{"key":"e_1_3_2_49_2","unstructured":"Jia-Chen Gu Hao-Xiang Xu Jun-Yu Ma Pan Lu Zhen-Hua Ling Kai-Wei Chang and Nanyun Peng. 2024. Model editing harms general abilities of large language models: Regularization to the rescue. arXiv preprint arXiv:2401.04700 (2024)."},{"key":"e_1_3_2_50_2","doi-asserted-by":"crossref","unstructured":"Anshita Gupta Debanjan Mondal Akshay Krishna Sheshadri Wenlong Zhao Xiang Lorraine Li Sarah Wiegreffe and Niket Tandon. 2023. Editing common sense in transformers. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2023.emnlp-main.511"},{"key":"e_1_3_2_51_2","doi-asserted-by":"crossref","unstructured":"Akshat Gupta Dev Sajnani and Gopala Anumanchipalli. 2024. A unified framework for model editing. arXiv preprint arXiv:2403.14236 (2024).","DOI":"10.18653\/v1\/2024.findings-emnlp.903"},{"key":"e_1_3_2_52_2","unstructured":"David Ha Andrew Dai and Quoc V. Le. 2016. HyperNetworks. arXiv:1609.09106. Retrieved from https:\/\/arxiv.org\/abs\/1609.09106"},{"key":"e_1_3_2_53_2","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","author":"Hartvigsen Thomas","year":"2023","unstructured":"Thomas Hartvigsen, Swami Sankaranarayanan, Hamid Palangi, Yoon Kim, and Marzyeh Ghassemi. 2023. Aging with GRACE: Lifelong model editing with discrete key-value adaptors. In Proceedings of the Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_54_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.eacl-main.199"},{"key":"e_1_3_2_55_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_56_2","volume-title":"More than a Feeling: Benchmarks for Sentiment Analysis Accuracy: Heitmann, Mark","author":"Heitmann Mark","year":"2020","unstructured":"Mark Heitmann. 2020. More than a feeling: Benchmarks for sentiment analysis accuracy. In More than a Feeling: Benchmarks for Sentiment Analysis Accuracy: Heitmann, Mark."},{"key":"e_1_3_2_57_2","article-title":"Image captioning: Transforming objects into words","author":"Herdade Simao","year":"2019","unstructured":"Simao Herdade, Armin Kappeler, Kofi Boakye, and Joao Soares. 2019. Image captioning: Transforming objects into words. Advances in Neural Information Processing Systems 32 (2019).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_58_2","unstructured":"Evan Hernandez Belinda Z. Li and Jacob Andreas. 2023. Inspecting and editing knowledge representations in language models. arXiv preprint arXiv:2304.00740 (2023)."},{"key":"e_1_3_2_59_2","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_3_2_60_2","doi-asserted-by":"crossref","unstructured":"Or Honovich Thomas Scialom Omer Levy and Timo Schick. 2022. Unnatural instructions: Tuning language models with (Almost) no human labor. In The 61st Annual Meeting Of The Association For Computational Linguistics.","DOI":"10.18653\/v1\/2023.acl-long.806"},{"key":"e_1_3_2_61_2","unstructured":"Timothy Hospedales Antreas Antoniou Paul Micaelli and Amos Storkey. 2021. Meta-learning in neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence 44 9 (2021) 5149\u20135169."},{"key":"e_1_3_2_62_2","doi-asserted-by":"crossref","unstructured":"Chenhui Hu Pengfei Cao Yubo Chen Kang Liu and Jun Zhao. 2024. Wilke: Wise-layer knowledge editor for lifelong knowledge editing. arXiv preprint arXiv:2402.10987 (2024).","DOI":"10.18653\/v1\/2024.findings-acl.207"},{"key":"e_1_3_2_63_2","unstructured":"Edward J. Hu Phillip Wallis Zeyuan Allen-Zhu Yuanzhi Li Shean Wang Lu Wang Weizhu Chen et\u00a0al. 2022. LoRA: Low-Rank adaptation of large language models. In International Conference on Learning Representations."},{"key":"e_1_3_2_64_2","unstructured":"Linmei Hu Zeyi Liu Ziwang Zhao Lei Hou Liqiang Nie and Juanzi Li. 2023. A survey of knowledge enhanced pre-trained language models. IEEE Transactions on Knowledge and Data Engineering (2023)."},{"key":"e_1_3_2_65_2","unstructured":"Zhiqiang Hu Lei Wang Yihuai Lan Wanyu Xu Ee-Peng Lim Lidong Bing Xing Xu Soujanya Poria and Roy Lee. 2023. LLM-Adapters: An adapter family for parameter-efficient fine-tuning of large language models. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing."},{"key":"e_1_3_2_66_2","unstructured":"Han Huang Haitian Zhong Qiang Liu Shu Wu Liang Wang and Tieniu Tan. 2024. KEBench: A benchmark on knowledge editing for large vision-language models. arXiv preprint arXiv:2403.07350 (2024)."},{"key":"e_1_3_2_67_2","unstructured":"Zeyu Huang Yikang Shen Xiaofeng Zhang Jie Zhou Wenge Rong and Zhang Xiong. 2023. Transformer-patcher: One mistake worth one neuron. In International Conference on Learning Representations."},{"key":"e_1_3_2_68_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-021-10004-4"},{"key":"e_1_3_2_69_2","unstructured":"Gabriel Ilharco Marco Tulio Ribeiro Mitchell Wortsman Ludwig Schmidt Hannaneh Hajishirzi and Ali Farhadi. 2023. Editing models with task arithmetic. In International Conference on Learning Representations."},{"key":"e_1_3_2_70_2","doi-asserted-by":"crossref","unstructured":"Yuxin Jiang Yufei Wang Chuhan Wu Wanjun Zhong Xingshan Zeng Jiahui Gao Liangyou Li Xin Jiang Lifeng Shang Ruiming Tang Qun Liu and Wei Wang. 2024. Learning to edit: Aligning LLMs with knowledge editing. arXiv preprint arXiv:2402.11905 (2024).","DOI":"10.18653\/v1\/2024.acl-long.258"},{"key":"e_1_3_2_71_2","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00324"},{"key":"e_1_3_2_72_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2021.103982"},{"key":"e_1_3_2_73_2","doi-asserted-by":"publisher","DOI":"10.1007\/s13347-023-00606-x"},{"key":"e_1_3_2_74_2","doi-asserted-by":"crossref","unstructured":"Enkelejda Kasneci Kathrin Sessler Stefan K\u00fcchemann Maria Bannert Daryna Dementieva Frank Fischer Urs Gasser Georg Groh Stephan G\u00fcnnemann Eyke H\u00fcllermeier Stephan Krusche Gitta Kutyniok Tilman Michaeli Claudia Nerdel J\u00fcrgen Pfeffer Oleksandra Poquet Michael Sailer Albrecht Schmidt Tina Seidel Matthias Stadler Jochen Weller Jochen Kuhn and Gjergji Kasneci. 2023. ChatGPT for good? On opportunities and challenges of large language models for education. Learning and Individual Differences 103 (2023) 102274.","DOI":"10.1016\/j.lindif.2023.102274"},{"key":"e_1_3_2_75_2","unstructured":"Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics."},{"key":"e_1_3_2_76_2","doi-asserted-by":"publisher","DOI":"10.1109\/TC.1972.5008975"},{"key":"e_1_3_2_77_2","doi-asserted-by":"crossref","unstructured":"Tom Kwiatkowski Jennimaria Palomaki Olivia Redfield Michael Collins Ankur Parikh Chris Alberti Danielle Epstein Illia Polosukhin Jacob Devlin Kenton Lee Kristina Toutanova Llion Jones Matthew Kelcey Ming-Wei Chang Andrew M. Dai Jakob Uszkoreit Quoc Le and Slav Petrov. 2019. Natural questions: A benchmark for question answering research. Transactions of the Association for Computational Linguistics 7 (2019) 453\u2013466.","DOI":"10.1162\/tacl_a_00276"},{"key":"e_1_3_2_78_2","unstructured":"Kyungjae Lee Wookje Han Seung won Hwang Hwaran Lee Joonsuk Park and Sang-Woo Lee. 2022. Plug-and-play adaptation for continuously-updated QA. In Proceedings of the Annual Meeting of the Association for Computational Linguistics."},{"key":"e_1_3_2_79_2","doi-asserted-by":"crossref","unstructured":"Omer Levy Minjoon Seo Eunsol Choi and Luke Zettlemoyer. 2017. Zero-shot relation extraction via reading comprehension. In Proceedings of the Conference on Computational Natural Language Learning 2017.","DOI":"10.18653\/v1\/K17-1034"},{"key":"e_1_3_2_80_2","doi-asserted-by":"crossref","unstructured":"Daliang Li Ankit Singh Rawat Manzil Zaheer Xin Wang Michal Lukasik Andreas Veit Felix X. Yu and Sanjiv Kumar. 2023. Large language models with controllable working memory. In Proceedings of the Annual Meeting of the Association for Computational Linguistics.","DOI":"10.18653\/v1\/2023.findings-acl.112"},{"key":"e_1_3_2_81_2","unstructured":"Junnan Li Dongxu Li Silvio Savarese and Steven Hoi. 2023. Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. In International Conference on Machine Learning."},{"key":"e_1_3_2_82_2","unstructured":"Shuaiyi Li Yang Deng Deng Cai Hongyuan Lu Liang Chen and Wai Lam. 2024. Consecutive model editing with batch alongside HooK layers. arXiv preprint arXiv:2403.05330 (2024)."},{"key":"e_1_3_2_83_2","unstructured":"Xiaopeng Li Shasha Li Bin Ji Shezheng Song Xi Wang Jun Ma Jie Yu Xiaodong Liu Jing Wang and Weimin Zhang. 2024. SWEA: Changing factual knowledge in large language models via subject word embedding altering. arXiv preprint arXiv:2401.17809 (2024)."},{"key":"e_1_3_2_84_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i17.29818"},{"key":"e_1_3_2_85_2","unstructured":"Xiaonan Li and Xipeng Qiu. 2023. Finding supporting examples for in-context learning. arXiv preprint arXiv:2302.13539 (2023)."},{"key":"e_1_3_2_86_2","unstructured":"Yuchao Li Fuli Luo Chuanqi Tan Mengdi Wang Songfang Huang Shen Li and Junjie Bai. 2022. Parameter-efficient sparsity for large language models fine-tuning. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence."},{"key":"e_1_3_2_87_2","unstructured":"Zhoubo Li Ningyu Zhang Yunzhi Yao Mengru Wang Xi Chen and Huajun Chen. 2024. Unveiling the pitfalls of knowledge editing for large language models. In International Conference on Learning Representations."},{"key":"e_1_3_2_88_2","doi-asserted-by":"crossref","unstructured":"Q. Vera Liao and Jennifer Wortman Vaughan. 2023. AI transparency in the age of LLMs: A human-centered research roadmap. arXiv preprint arXiv:2306.01941 (2023).","DOI":"10.1162\/99608f92.8036d03b"},{"key":"e_1_3_2_89_2","unstructured":"Hao Liu Carmelo Sferrazza and Pieter Abbeel. 2023. Chain of hindsight aligns language models with feedback. In International Conference on Learning Representations."},{"key":"e_1_3_2_90_2","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","author":"Liu Haokun","year":"2022","unstructured":"Haokun Liu, Derek Tam, Mohammed Muqeeth, Jay Mohta, Tenghao Huang, Mohit Bansal, and Colin A. Raffel. 2022. Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning. In Proceedings of the Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_91_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-21348-0_30"},{"key":"e_1_3_2_92_2","unstructured":"Yinhan Liu Myle Ott Naman Goyal Jingfei Du Mandar Joshi Danqi Chen Omer Levy Mike Lewis Luke Zettlemoyer and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_93_2","unstructured":"Yun Luo Zhen Yang Fandong Meng Yafu Li Jie Zhou and Yue Zhang. 2023. An empirical study of catastrophic forgetting in large language models during continual fine-tuning. arXiv preprint arXiv:2308.08747 (2023)."},{"key":"e_1_3_2_94_2","unstructured":"Jun-Yu Ma Jia-Chen Gu Zhen-Hua Ling Quan Liu and Cong Liu. 2023. Untying the reversal curse via bidirectional language model editing. arXiv preprint arXiv:2310.10322 (2023)."},{"key":"e_1_3_2_95_2","unstructured":"Yuxuan Ma. 2021. distilgpt2-finetuned-wikitext2. Retrieved November 2 2023 from https:\/\/huggingface.co\/MYX4567\/distilgpt2-finetuned-wikitext2"},{"key":"e_1_3_2_96_2","doi-asserted-by":"crossref","unstructured":"Aman Madaan Niket Tandon Peter Clark and Yiming Yang. 2022. Memory-assisted prompt editing to improve GPT-3 after deployment. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2022.emnlp-main.183"},{"key":"e_1_3_2_97_2","doi-asserted-by":"crossref","unstructured":"Potsawee Manakul Adian Liusie and Mark Gales. 2023. SelfCheckGPT: Zero-resource black-box hallucination detection for generative large language models. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2023.emnlp-main.557"},{"key":"e_1_3_2_98_2","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","author":"Meng Kevin","year":"2022","unstructured":"Kevin Meng, David Bau, Alex Andonian, and Yonatan Belinkov. 2022. Locating and editing factual associations in GPT. In Proceedings of the Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_99_2","unstructured":"Kevin Meng Arnab Sen Sharma Alex J. Andonian Yonatan Belinkov and David Bau. 2023. Mass-editing memory in a transformer. In International Conference on Learning Representations."},{"key":"e_1_3_2_100_2","unstructured":"Jacob Menick Maja Trebacz Vladimir Mikulik John Aslanides Francis Song Martin Chadwick Mia Glaese Susannah Young Lucy Campbell-Gillingham Geoffrey Irving and Nat McAleese. 2022. Teaching language models to support answers with verified quotes. arXiv preprint arXiv:2203.11147 (2022)."},{"key":"e_1_3_2_101_2","doi-asserted-by":"publisher","DOI":"10.1145\/3605943"},{"key":"e_1_3_2_102_2","unstructured":"Eric Mitchell Charles Lin Antoine Bosselut Chelsea Finn and Christopher D. Manning. 2022. Fast model editing at scale. In Proceedings of the International Conference on Machine Learning."},{"key":"e_1_3_2_103_2","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Mitchell Eric","year":"2022","unstructured":"Eric Mitchell, Charles Lin, Antoine Bosselut, Christopher D. Manning, and Chelsea Finn. 2022. Memory-based model editing at scale. In Proceedings of the International Conference on Machine Learning."},{"key":"e_1_3_2_104_2","doi-asserted-by":"crossref","unstructured":"Niklas Muennighoff Thomas Wang Lintang Sutawika Adam Roberts Stella Biderman Teven Le Scao M. Saiful Bari Sheng Shen Zheng-Xin Yong Hailey Schoelkopf Xiangru Tang Dragomir Radev Alham Fikri Aji Khalid Almubarak Samuel Albanie Zaid Alyafeai Albert Webson Edward Raff and Colin Raffel. 2023. Crosslingual generalization through multitask finetuning. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers).","DOI":"10.18653\/v1\/2023.acl-long.891"},{"key":"e_1_3_2_105_2","doi-asserted-by":"crossref","unstructured":"Shikhar Murty Christopher D. Manning Scott M. Lundberg and Marco T\u00falio Ribeiro. 2022. Fixing model bugs with natural language patches. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2022.emnlp-main.797"},{"key":"e_1_3_2_106_2","unstructured":"Thanh Tam Nguyen Thanh Trung Huynh Phi Le Nguyen Alan Wee-Chung Liew Hongzhi Yin and Quoc Viet Hung Nguyen. 2022. A survey of machine unlearning. arXiv preprint arXiv:2209.02299 (2022)."},{"key":"e_1_3_2_107_2","unstructured":"Shiwen Ni Dingwei Chen Chengming Li Xiping Hu Ruifeng Xu and Min Yang. 2023. Forgetting before Learning: Utilizing parametric arithmetic for knowledge updating in large language models. arXiv preprint arXiv:2311.08011 (2023)."},{"key":"e_1_3_2_108_2","doi-asserted-by":"crossref","unstructured":"Yasumasa Onoe Michael Zhang Eunsol Choi and Greg Durrett. 2022. Entity cloze by date: What LMs know about unseen entities. In Findings of Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics.","DOI":"10.18653\/v1\/2022.findings-naacl.52"},{"key":"e_1_3_2_109_2","doi-asserted-by":"crossref","unstructured":"Yasumasa Onoe Michael J. Q. Zhang Shankar Padmanabhan Greg Durrett and Eunsol Choi. 2023. Can LMs learn new entities from descriptions? Challenges in propagating injected knowledge. In Proceedings of the Annual Meeting of the Association for Computational Linguistics.","DOI":"10.18653\/v1\/2023.acl-long.300"},{"key":"e_1_3_2_110_2","unstructured":"OpenAI. 2023. GPT-4 Technical Report. arXiv preprint arXiv:2303.08774 (2023)."},{"key":"e_1_3_2_111_2","doi-asserted-by":"crossref","unstructured":"Hariom A. Pandya and Brijesh S. Bhatt. 2021. Question Answering Survey: Directions Challenges Datasets Evaluation Matrices. arXiv preprint arXiv:2112.03572 (2021).","DOI":"10.20944\/preprints202112.0136.v1"},{"key":"e_1_3_2_112_2","unstructured":"Baolin Peng Michel Galley Pengcheng He Hao Cheng Yujia Xie Yu Hu Qiuyuan Huang Lars Liden Zhou Yu Weizhu Chen and Jianfeng Gao. 2023. Check your facts and try again: Improving large language models with external knowledge and automated feedback. arXiv preprint arXiv:2302.12813 (2023)."},{"key":"e_1_3_2_113_2","unstructured":"Baolin Peng Chunyuan Li Pengcheng He Michel Galley and Jianfeng Gao. 2023. Instruction tuning with GPT-4. arXiv preprint arXiv:2304.03277 (2023)."},{"key":"e_1_3_2_114_2","doi-asserted-by":"crossref","unstructured":"Ethan Perez Saffron Huang Francis Song Trevor Cai Roman Ring John Aslanides Amelia Glaese Nat McAleese and Geoffrey Irving. 2022. Red teaming language models with language models. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2022.emnlp-main.225"},{"key":"e_1_3_2_115_2","doi-asserted-by":"crossref","unstructured":"Fabio Petroni Aleksandra Piktus Angela Fan Patrick Lewis Majid Yazdani Nicola De Cao James Thorne Yacine Jernite Vladimir Karpukhin Jean Maillard Vassilis Plachouras Tim Rockt\u00e4schel and Sebastian Riedel. 2021. KILT: A benchmark for knowledge intensive language tasks. In Proceedings of the Annual Meeting of the Association for Computational Linguistics.","DOI":"10.18653\/v1\/2021.naacl-main.200"},{"key":"e_1_3_2_116_2","doi-asserted-by":"crossref","unstructured":"Fabio Petroni Tim Rockt\u00e4schel Sebastian Riedel Patrick Lewis Anton Bakhtin Yuxiang Wu and Alexander Miller. 2019. Language models as knowledge bases? In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/D19-1250"},{"key":"e_1_3_2_117_2","doi-asserted-by":"crossref","unstructured":"Yuval Pinter and Michael Elhadad. 2023. Emptying the Ocean with a Spoon: Should we edit models? In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/2023.findings-emnlp.1012"},{"key":"e_1_3_2_118_2","unstructured":"Yujia Qin Xiaozhi Wang Yusheng Su Yankai Lin Ning Ding Jing Yi Weize Chen Zhiyuan Liu Juanzi Li Lei Hou Peng Li Maosong Sun and Jie Zhou. 2022. Exploring Universal Intrinsic Task Subspace via Prompt Tuning. arXiv preprint arXiv:2110.07867 (2022)."},{"key":"e_1_3_2_119_2","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. OpenAI (2018)."},{"key":"e_1_3_2_120_2","unstructured":"Colin Raffel Noam Shazeer Adam Roberts Katherine Lee Sharan Narang Michael Matena Yanqi Zhou Wei Li and Peter J. Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research 21 140 (2020) 1\u201367."},{"key":"e_1_3_2_121_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Ravi Sachin","year":"2016","unstructured":"Sachin Ravi and Hugo Larochelle. 2016. Optimization as a model for few-shot learning. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_122_2","doi-asserted-by":"crossref","unstructured":"Nils Reimers and Iryna Gurevych. 2019. Sentence-BERT: Sentence embeddings using siamese BERT-networks. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/D19-1410"},{"key":"e_1_3_2_123_2","doi-asserted-by":"publisher","DOI":"10.1017\/S1351324997001502"},{"key":"e_1_3_2_124_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.230"},{"key":"e_1_3_2_125_2","doi-asserted-by":"crossref","unstructured":"Stephen Roller Emily Dinan Naman Goyal Da Ju Mary Williamson Yinhan Liu Jing Xu Myle Ott Kurt Shuster Eric M. Smith Y-Lan Boureau and Jason Weston. 2021. Recipes for building an open-domain chatbot. In Proceedings of the Conference of the European Chapter of the Association for Computational Linguistics.","DOI":"10.18653\/v1\/2021.eacl-main.24"},{"key":"e_1_3_2_126_2","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","author":"Santurkar Shibani","year":"2021","unstructured":"Shibani Santurkar, Dimitris Tsipras, Mahalaxmi Elango, David Bau, Antonio Torralba, and Aleksander Madry. 2021. Editing a classifier by rewriting its prediction rules. In Proceedings of the Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_127_2","unstructured":"Christoph Schuhmann Robert Kaczmarczyk Aran Komatsuzaki Aarush Katta Richard Vencu Romain Beaumont Jenia Jitsev Theo Coombes and Clayton Mullis. 2021. LAION-400M: Open dataset of CLIP-Filtered 400 Million image-text pairs. In Proceedings of the Advances in Neural Information Processing Systems Workshop Datacentric AI."},{"key":"e_1_3_2_128_2","doi-asserted-by":"crossref","unstructured":"Tal Schuster Adam Fisch and Regina Barzilay. 2021. Get your Vitamin C! Robust fact verification with contrastive evidence. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics.","DOI":"10.18653\/v1\/2021.naacl-main.52"},{"key":"e_1_3_2_129_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.410"},{"key":"e_1_3_2_130_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.osnem.2020.100104"},{"key":"e_1_3_2_131_2","unstructured":"Arnab Sen Sharma David Atkinson and David Bau. 2024. Locating and editing factual associations in mamba. arXiv:2404.03646. Retrieved from https:\/\/arxiv.org\/abs\/2404.03646"},{"key":"e_1_3_2_132_2","unstructured":"Yucheng Shi Qiaoyu Tan Xuansheng Wu Shaochen Zhong Kaixiong Zhou and Ninghao Liu. 2024. Retrieval-enhanced knowledge editing for multi-hop question answering in language models. arXiv:2403.19631. Retrieved from https:\/\/arxiv.org\/abs\/2403.19631"},{"key":"e_1_3_2_133_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.346"},{"key":"e_1_3_2_134_2","volume-title":"ICLR","author":"Sinitsin Anton","year":"2020","unstructured":"Anton Sinitsin, Vsevolod Plokhotnyuk, Dmitry Pyrkin, Sergei Popov, and Artem Babenko. 2020. Editable neural networks. In ICLR."},{"key":"e_1_3_2_135_2","unstructured":"Chenyang Song Xu Han Zheni Zeng Kuai Li Chen Chen Zhiyuan Liu Maosong Sun and Tao Yang. 2023. ConPET: Continual parameter-efficient tuning for large language models. arXiv:2309.14763. Retrieved from https:\/\/arxiv.org\/abs\/2309.14763"},{"key":"e_1_3_2_136_2","unstructured":"Feifan Song Bowen Yu Minghao Li Haiyang Yu Fei Huang Yongbin Li and Houfeng Wang. 2023. Preference ranking optimization for human alignment. arXiv:2306.17492. Retrieved from https:\/\/arxiv.org\/abs\/2306.17492"},{"key":"e_1_3_2_137_2","unstructured":"Xiaoshuai Song Zhengyang Wang Keqing He Guanting Dong Jinxu Zhao and Weiran Xu. 2024. Knowledge editing on black-box large language models. arXiv:2402.08631. Retrieved from https:\/\/arxiv.org\/abs\/2402.08631"},{"key":"e_1_3_2_138_2","doi-asserted-by":"publisher","DOI":"10.1613\/jair.1.12007"},{"key":"e_1_3_2_139_2","unstructured":"Hongjin Su Jungo Kasai Chen Henry Wu Weijia Shi Tianlu Wang Jiayi Xin Rui Zhang Mari Ostendorf Luke Zettlemoyer Noah A Smith and Tao Yu. 2022. Selective annotation makes language models better few-shot learners. arXiv preprint arXiv:2209.01975 (2022)."},{"key":"e_1_3_2_140_2","unstructured":"Alon Talmor Jonathan Herzig Nicholas Lourie and Jonathan Berant. 2018. Commonsenseqa: A question answering challenge targeting commonsense knowledge. arXiv:1811.00937. Retrieved from https:\/\/arxiv.org\/abs\/1811.00937"},{"key":"e_1_3_2_141_2","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","author":"Tanno Ryutaro","year":"2022","unstructured":"Ryutaro Tanno, Melanie F. Pradier, Aditya Nori, and Yingzhen Li. 2022. Repairing neural networks by leaving the right past behind. In Proceedings of the Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_142_2","unstructured":"Rohan Taori Ishaan Gulrajani Tianyi Zhang Yann Dubois Xuechen Li Carlos Guestrin Percy Liang and Tatsunori B. Hashimoto. 2023. Stanford Alpaca: An Instruction-following LLaMA Model. Retrieved November 15 2023 from https:\/\/github.com\/tatsu-lab\/stanford_alpaca"},{"key":"e_1_3_2_143_2","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-023-02448-8"},{"key":"e_1_3_2_144_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1074"},{"key":"e_1_3_2_145_2","unstructured":"Hugo Touvron Thibaut Lavril Gautier Izacard Xavier Martinet Marie-Anne Lachaux Timoth\u00e9e Lacroix Baptiste Rozi\u00e8re Naman Goyal Eric Hambro Faisal Azhar Aurelien Rodriguez Armand Joulin Edouard Grave and Guillaume Lample. 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_146_2","unstructured":"Joaquin Vanschoren. 2018. Meta-learning: A survey. arXiv:1810.03548. Retrieved from https:\/\/arxiv.org\/abs\/1810.03548"},{"key":"e_1_3_2_147_2","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N. Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. Proceedings of the Advances in Neural Information Processing Systems (2017)."},{"key":"e_1_3_2_148_2","unstructured":"Johannes von Oswald Christian Henning Benjamin F. Grewe and Jo\u00e3o Sacramento. 2022. Continual Learning with Hypernetworks. arXiv:1906.00695. Retrieved from https:\/\/arxiv.org\/abs\/1906.00695"},{"key":"e_1_3_2_149_2","doi-asserted-by":"crossref","unstructured":"Denny Vrande\u010di\u0107 and Markus Kr\u00f6tzsch. 2014. Wikidata: A free collaborative knowledgebase. Commun. ACM 57 10 (2014) 78\u201385.","DOI":"10.1145\/2629489"},{"key":"e_1_3_2_150_2","unstructured":"Mengru Wang Ningyu Zhang Ziwen Xu Zekun Xi Shumin Deng Yunzhi Yao Qishen Zhang Linyi Yang Jindong Wang and Huajun Chen. 2024. Detoxifying large language models via knowledge editing. arXiv:2403.14472. Retrieved from https:\/\/arxiv.org\/abs\/2403.14472"},{"key":"e_1_3_2_151_2","unstructured":"Peiyi Wang Lei Li Liang Chen Dawei Zhu Binghuai Lin Yunbo Cao Qi Liu Tianyu Liu and Zhifang Sui. 2023. Large language models are not fair evaluators. arXiv:2305.17926. Retrieved from https:\/\/arxiv.org\/abs\/2305.17926"},{"key":"e_1_3_2_152_2","unstructured":"Peng Wang Zexi Li Ningyu Zhang Ziwen Xu Yunzhi Yao Yong Jiang Pengjun Xie Fei Huang and Huajun Chen. 2024. WISE: Rethinking the knowledge memory for lifelong model editing of large language models. arXiv:2405.14768. Retrieved from https:\/\/arxiv.org\/abs\/2405.14768"},{"key":"e_1_3_2_153_2","unstructured":"Peng Wang Ningyu Zhang Xin Xie Yunzhi Yao Bozhong Tian Mengru Wang Zekun Xi Siyuan Cheng Kangwei Liu Guozhou Zheng and Huajun Chen. 2023. EasyEdit: An easy-to-use knowledge editing framework for large language models. arXiv preprint arXiv:2308.07269 (2023)."},{"key":"e_1_3_2_154_2","doi-asserted-by":"crossref","unstructured":"Ruize Wang Duyu Tang Nan Duan Zhongyu Wei Xuan-Jing Huang Jianshu Ji Guihong Cao Daxin Jiang and Ming Zhou. 2021. K-Adapter: Infusing knowledge into pre-trained models with adapters. In Findings of the Association for Computational Linguistics. 1405\u20131418.","DOI":"10.18653\/v1\/2021.findings-acl.121"},{"key":"e_1_3_2_155_2","unstructured":"Weixuan Wang Barry Haddow and Alexandra Birch. 2023. Retrieval-augmented multilingual knowledge editing. arXiv:2312.13040. Retrieved from https:\/\/arxiv.org\/abs\/2312.13040"},{"key":"e_1_3_2_156_2","unstructured":"Yiwei Wang Muhao Chen Nanyun Peng and Kai-Wei Chang. 2024. Deepedit: Knowledge editing as decoding with constraints. arXiv:2401.10471. Retrieved from https:\/\/arxiv.org\/abs\/2401.10471"},{"key":"e_1_3_2_157_2","unstructured":"Yu Wang Xiusi Chen Jingbo Shang and Julian McAuley. 2024. MemoryLLM: Towards self-updatable large language models. arXiv:2402.04624. Retrieved from https:\/\/arxiv.org\/abs\/2402.04624"},{"key":"e_1_3_2_158_2","doi-asserted-by":"crossref","unstructured":"Yizhong Wang Yeganeh Kordi Swaroop Mishra Alisa Liu Noah A. Smith Daniel Khashabi and Hannaneh Hajishirzi. 2022. Self-Instruct: Aligning language model with self generated instructions. arXiv:2212.10560. Retrieved from https:\/\/arxiv.org\/abs\/2212.10560","DOI":"10.18653\/v1\/2023.acl-long.754"},{"key":"e_1_3_2_159_2","doi-asserted-by":"crossref","unstructured":"Yaqing Wang Subhabrata Mukherjee Xiaodong Liu Jing Gao Ahmed Hassan Awadallah and Jianfeng Gao. 2022. Adamix: Mixture-of-adapter for parameter-efficient tuning of large language models. arXiv:2205.12410. Retrieved from https:\/\/arxiv.org\/abs\/2205.12410","DOI":"10.18653\/v1\/2022.emnlp-main.388"},{"key":"e_1_3_2_160_2","unstructured":"Yufei Wang Wanjun Zhong Liangyou Li Fei Mi Xingshan Zeng Wenyong Huang Lifeng Shang Xin Jiang and Qun Liu. 2023. Aligning large language models with human: A survey. arXiv:2307.12966. Retrieved from https:\/\/arxiv.org\/abs\/2307.12966"},{"key":"e_1_3_2_161_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-022-10144-1"},{"key":"e_1_3_2_162_2","unstructured":"Jason Wei Maarten Bosma Vincent Zhao Kelvin Guu Adams Wei Yu Brian Lester Nan Du Andrew M. Dai and Quoc V. Le. 2021. Finetuned Language Models are Zero-Shot Learners. In International Conference on Learning Representations."},{"key":"e_1_3_2_163_2","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V. Le, and Denny Zhou. 2022. Chain-of-thought prompting elicits reasoning in large language models. In Proceedings of the Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_164_2","unstructured":"Zihao Wei Jingcheng Deng Liang Pang Hanxing Ding Huawei Shen and Xueqi Cheng. 2024. Mlake: Multilingual knowledge editing benchmark for large language models. arXiv:2404.04990. Retrieved from https:\/\/arxiv.org\/abs\/2404.04990"},{"key":"e_1_3_2_165_2","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","year":"2022","unstructured":"Mitchell Wortsman, Gabriel Ilharco, Jong Wook Kim, Mike Li, Simon Kornblith, Rebecca Roelofs, Raphael Gontijo Lopes, Hannaneh Hajishirzi, Ali Farhadi, Hongseok Namkoong, and Ludwig Schmidt. 2022. Robust fine-tuning of zero-shot models. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition."},{"key":"e_1_3_2_166_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.174"},{"key":"e_1_3_2_167_2","unstructured":"Xiaobao Wu Liangming Pan William Yang Wang and Anh Tuan Luu. 2024. Updating language models with unstructured facts: Towards practical knowledge editing. arXiv:2402.18909. Retrieved from https:\/\/arxiv.org\/abs\/2402.18909"},{"key":"e_1_3_2_168_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.632"},{"key":"e_1_3_2_169_2","unstructured":"Junsang Yoon Akshat Gupta and Gopala Anumanchipalli. 2024. Is bigger edit batch size always better?\u2013An empirical study on model editing with Llama-3. arXiv:2405.00664. Retrieved from https:\/\/arxiv.org\/abs\/2405.00664"},{"key":"e_1_3_2_170_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i17.29916"},{"key":"e_1_3_2_171_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-short.1"},{"key":"e_1_3_2_172_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.586"},{"key":"e_1_3_2_173_2","doi-asserted-by":"crossref","unstructured":"Ningyu Zhang Yunzhi Yao Bozhong Tian PengWang Shumin Deng MengruWang Zekun Xi Shengyu Mao Jintian Zhang Yuansheng Ni Siyuan Cheng Ziwen Xu Xin Xu Jia-Chen Gu Yong Jiang Pengjun Xie Fei Huang Lei Liang Zhiqiang Zhang Xiaowei Zhu Jun Zhou and Huajun Chen. 2024. A comprehensive study of knowledge editing for large language models. arXiv preprint arXiv:2401.01286 (2024).","DOI":"10.18653\/v1\/2024.findings-emnlp.40"},{"key":"e_1_3_2_174_2","unstructured":"Wayne Xin Zhao Kun Zhou Junyi Li Tianyi Tang Xiaolei Wang Yupeng Hou Yingqian Min Beichen Zhang Junjie Zhang Zican Dong Yifan Du Chen Yang Yushuo Chen Zhipeng Chen Jinhao Jiang Ruiyang Ren Yifan Li Xinyu Tang Zikang Liu Peiyu Liu Jian-Yun Nie and Ji-Rong Wen. 2023. A survey of large language models. arXiv preprint arXiv:2303.18223 (2023)."},{"key":"e_1_3_2_175_2","doi-asserted-by":"crossref","unstructured":"Ce Zheng Lei Li Qingxiu Dong Yuxuan Fan Zhiyong Wu Jingjing Xu and Baobao Chang. 2023. Can We Edit Factual Knowledge by In-Context Learning? arXiv:2305.12740. Retrieved from https:\/\/arxiv.org\/abs\/2305.12740","DOI":"10.18653\/v1\/2023.emnlp-main.296"},{"key":"e_1_3_2_176_2","doi-asserted-by":"crossref","unstructured":"Zexuan Zhong Zhengxuan Wu Christopher D. Manning Christopher Potts and Danqi Chen. 2023. MQuAKE: Assessing knowledge editing in language models via multi-hop questions. arXiv:2305.14795. Retrieved from https:\/\/arxiv.org\/abs\/2305.14795","DOI":"10.18653\/v1\/2023.emnlp-main.971"},{"key":"e_1_3_2_177_2","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581318"},{"key":"e_1_3_2_178_2","unstructured":"Chen Zhu Ankit Singh Rawat Manzil Zaheer Srinadh Bhojanapalli Daliang Li Felix Yu and Sanjiv Kumar. 2020. Modifying Memories in Transformer Models. arXiv:2012.00363. Retrieved from https:\/\/arxiv.org\/abs\/2012.00363"},{"key":"e_1_3_2_179_2","article-title":"A comprehensive survey on transfer learning","author":"Zhuang Fuzhen","year":"2020","unstructured":"Fuzhen Zhuang, Zhiyuan Qi, Keyu Duan, Dongbo Xi, Yongchun Zhu, Hengshu Zhu, Hui Xiong, and Qing He. 2020. A comprehensive survey on transfer learning. Proc. IEEE 109, 1 (2020), 43\u201376.","journal-title":"Proc. IEEE"},{"key":"e_1_3_2_180_2","unstructured":"Daniel M. Ziegler Nisan Stiennon Jeffrey Wu Tom B. Brown Alec Radford Dario Amodei Paul Christiano and Geoffrey Irving. 2019. Fine-tuning language models from human preferences. arXiv:1909.08593. Retrieved from https:\/\/arxiv.org\/abs\/1909.08593"}],"container-title":["ACM Computing Surveys"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3698590","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3698590","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:09:44Z","timestamp":1750295384000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3698590"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,11]]},"references-count":179,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,3,31]]}},"alternative-id":["10.1145\/3698590"],"URL":"https:\/\/doi.org\/10.1145\/3698590","relation":{},"ISSN":["0360-0300","1557-7341"],"issn-type":[{"value":"0360-0300","type":"print"},{"value":"1557-7341","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,11]]},"assertion":[{"value":"2023-12-14","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-09-16","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-11-11","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}