{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T13:38:59Z","timestamp":1770730739509,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":25,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,11,19]],"date-time":"2023-11-19T00:00:00Z","timestamp":1700352000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"National Natural Science Foundation of China","award":["62402273"],"award-info":[{"award-number":["62402273"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,11,19]]},"DOI":"10.1145\/3689217.3690614","type":"proceedings-article","created":{"date-parts":[[2024,11,19]],"date-time":"2024-11-19T18:22:01Z","timestamp":1732040521000},"page":"69-76","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Have You Merged My Model? On The Robustness of Large Language Model IP Protection Methods Against Model Merging"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3189-8223","authenticated-orcid":false,"given":"Tianshuo","family":"Cong","sequence":"first","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-2630-2680","authenticated-orcid":false,"given":"Delong","family":"Ran","sequence":"additional","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-0776-8839","authenticated-orcid":false,"given":"Zesen","family":"Liu","sequence":"additional","affiliation":[{"name":"Xidian University, Xi'an, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-3879-9080","authenticated-orcid":false,"given":"Xinlei","family":"He","sequence":"additional","affiliation":[{"name":"The Hong Kong University of Science and Technology (Guangzhou), Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-6346-5240","authenticated-orcid":false,"given":"Jinyuan","family":"Liu","sequence":"additional","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-0565-5035","authenticated-orcid":false,"given":"Yichen","family":"Gong","sequence":"additional","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8776-8730","authenticated-orcid":false,"given":"Qi","family":"Li","sequence":"additional","affiliation":[{"name":"Tsinghua University &amp; Zhongguancun Laboratory, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1086-0288","authenticated-orcid":false,"given":"Anyu","family":"Wang","sequence":"additional","affiliation":[{"name":"Tsinghua University &amp; Zhongguancun Laboratory, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7846-269X","authenticated-orcid":false,"given":"Xiaoyun","family":"Wang","sequence":"additional","affiliation":[{"name":"Tsinghua University &amp; Zhongguancun Laboratory &amp; National Financial Cryptography Research Center, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2024,11,19]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"1615","volume-title":"27th USENIX Security Symposium (USENIX Security)","author":"Adi Yossi","year":"2018","unstructured":"Yossi Adi, Carsten Baum, Moustapha Cisse, Benny Pinkas, and Joseph Keshet. Turning your weakness into a strength: Watermarking deep neural networks by backdooring. In 27th USENIX Security Symposium (USENIX Security), pages 1615--1631, 2018."},{"key":"e_1_3_2_1_2_1","volume-title":"Heres a free lunch: Sanitizing backdoored models with model merge. CoRR abs\/2402.19334","author":"Arora Ansh","year":"2024","unstructured":"Ansh Arora, Xuanli He, Maximilian Mozes, Srinibas Swain, Mark Dras, and Qiongkai Xu. Heres a free lunch: Sanitizing backdoored models with model merge. CoRR abs\/2402.19334, 2024."},{"key":"e_1_3_2_1_3_1","volume-title":"Do Duc Anh, and Soujanya Poria. Language models are homer simpson! safety re-alignment of fine-tuned language models through task arithmetic. CoRR abs\/2402.11746","author":"Bhardwaj Rishabh","year":"2024","unstructured":"Rishabh Bhardwaj, Do Duc Anh, and Soujanya Poria. Language models are homer simpson! safety re-alignment of fine-tuned language models through task arithmetic. CoRR abs\/2402.11746, 2024."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3433210.3437526"},{"key":"e_1_3_2_1_5_1","volume-title":"Thirty-seventh Conference on Neural Information Processing Systems (NeurIPS)","author":"Carlini Nicholas","year":"2023","unstructured":"Nicholas Carlini, Milad Nasr, Christopher A Choquette-Choo, Matthew Jagielski, Irena Gao, Pang Wei Koh, Daphne Ippolito, Florian Tram\u00e8r, and Ludwig Schmidt. Are aligned neural networks adversarially aligned? In Thirty-seventh Conference on Neural Information Processing Systems (NeurIPS), 2023."},{"key":"e_1_3_2_1_6_1","volume-title":"Evaluating the feasibility of chatgpt in healthcare: an analysis of multiple clinical and research scenarios. Journal of medical systems, 47(1):33","author":"Cascella Marco","year":"2023","unstructured":"Marco Cascella, Jonathan Montomoli, Valentina Bellini, and Elena Bignami. Evaluating the feasibility of chatgpt in healthcare: an analysis of multiple clinical and research scenarios. Journal of medical systems, 47(1):33, 2023."},{"key":"e_1_3_2_1_7_1","volume-title":"Training verifiers to solve math word problems. CoRR abs\/2110.14168","author":"Cobbe Karl","year":"2021","unstructured":"Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, Mark Chen, Heewoo Jun, Lukasz Kaiser, Matthias Plappert, Jerry Tworek, Jacob Hilton, Reiichiro Nakano, Christopher Hesse, and John Schulman. Training verifiers to solve math word problems. CoRR abs\/2110.14168, 2021."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3548606.3559355"},{"key":"e_1_3_2_1_9_1","first-page":"1206936","volume-title":"Frontiers in Education","author":"Dempere Juan","unstructured":"Juan Dempere, Kennedy Modugu, Allam Hesham, and Lakshmana Kumar Ramasamy. The impact of chatgpt on higher education. In Frontiers in Education, volume 8, page 1206936. Frontiers, 2023."},{"key":"e_1_3_2_1_10_1","volume-title":"Arcees mergekit: A toolkit for merging large language models. CoRR abs\/2403.13257","author":"Goddard Charles","year":"2024","unstructured":"Charles Goddard, Shamane Siriwardhana, Malikeh Ehghaghi, Luke Meyers, Vlad Karpukhin, Brian Benedict, Mark McQuade, and Jacob Solawetz. Arcees mergekit: A toolkit for merging large language models. CoRR abs\/2403.13257, 2024."},{"key":"e_1_3_2_1_11_1","volume-title":"International Conference on Learning Representations (ICLR)","author":"Hu Edward J","year":"2021","unstructured":"Edward J Hu, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al. Lora: Low-rank adaptation of large language models. In International Conference on Learning Representations (ICLR), 2021."},{"key":"e_1_3_2_1_12_1","volume-title":"The Eleventh International Conference on Learning Representations (ICLR)","author":"Ilharco Gabriel","year":"2022","unstructured":"Gabriel Ilharco, Marco Tulio Ribeiro, Mitchell Wortsman, Ludwig Schmidt, Hannaneh Hajishirzi, and Ali Farhadi. Editing models with task arithmetic. In The Eleventh International Conference on Learning Representations (ICLR), 2022."},{"key":"e_1_3_2_1_13_1","first-page":"17061","volume-title":"International Conference on Machine Learning (ICML)","author":"Kirchenbauer John","year":"2023","unstructured":"John Kirchenbauer, Jonas Geiping, YuxinWen, Jonathan Katz, Ian Miers, and Tom Goldstein. A watermark for large language models. In International Conference on Machine Learning (ICML), pages 17061--17084. PMLR, 2023."},{"key":"e_1_3_2_1_14_1","volume-title":"Watermarking llms with weight quantization. CoRR abs\/2310.11237","author":"Li Linyang","year":"2023","unstructured":"Linyang Li, Botian Jiang, Pengyu Wang, Ke Ren, Hang Yan, and Xipeng Qiu. Watermarking llms with weight quantization. CoRR abs\/2310.11237, 2023."},{"key":"e_1_3_2_1_15_1","volume-title":"AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models. CoRR abs\/2310.04451","author":"Liu Xiaogeng","year":"2023","unstructured":"Xiaogeng Liu, Nan Xu, Muhao Chen, and Chaowei Xiao. AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models. CoRR abs\/2310.04451, 2023."},{"key":"e_1_3_2_1_16_1","volume-title":"Wizardmath: Empowering mathematical reasoning for large language models via reinforced evol-instruct. CoRR abs\/2308.09583","author":"Luo Haipeng","year":"2023","unstructured":"Haipeng Luo, Qingfeng Sun, Can Xu, Pu Zhao, Jianguang Lou, Chongyang Tao, Xiubo Geng, Qingwei Lin, Shifeng Chen, and Dongmei Zhang. Wizardmath: Empowering mathematical reasoning for large language models via reinforced evol-instruct. CoRR abs\/2308.09583, 2023."},{"key":"e_1_3_2_1_17_1","volume-title":"A strongreject for empty jailbreaks. CoRR abs\/2402.10260","author":"Souly Alexandra","year":"2024","unstructured":"Alexandra Souly, Qingyuan Lu, Dillon Bowen, Tu Trinh, Elvis Hsieh, Sana Pandey, Pieter Abbeel, Justin Svegliato, Scott Emmons, Olivia Watkins, and Sam Toyer. A strongreject for empty jailbreaks. CoRR abs\/2402.10260, 2024."},{"key":"e_1_3_2_1_18_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale Dan Bikel Lukas Blecher Cristian Canton Ferrer Moya Chen Guillem Cucurull David Esiobu Jude Fernandes Jeremy Fu Wenyin Fu Brian Fuller Cynthia Gao Vedanuj Goswami Naman Goyal Anthony Hartshorn Saghar Hosseini Rui Hou Hakan Inan Marcin Kardas Viktor Kerkez Madian Khabsa Isabel Kloumann Artem Korenev Punit Singh Koura Marie-Anne Lachaux Thibaut Lavril Jenya Lee Diana Liskovich Yinghai Lu Yuning Mao Xavier Martinet Todor Mihaylov Pushkar Mishra Igor Molybog Yixin Nie Andrew Poulton Jeremy Reizenstein Rashi Rungta Kalyan Saladi Alan Schelten Ruan Silva Eric Michael Smith Ranjan Subramanian Xiaoqing Ellen Tan Binh Tang Ross Taylor Adina Williams Jian Xiang Kuan Puxin Xu Zheng Yan Iliyan Zarov Yuchen Zhang Angela Fan Melanie Kambadur Sharan Narang Aurelien Rodriguez Robert Stojnic Sergey Edunov and Thomas Scialom. Llama 2: Open foundation and fine-tuned chat models. CoRR abs\/2307.09288 2023."},{"key":"e_1_3_2_1_19_1","volume-title":"Chatgpt as your vehicle co-pilot: An initial attempt","author":"Wang Shiyi","year":"2023","unstructured":"Shiyi Wang, Yuxuan Zhu, Zhiheng Li, Yutong Wang, Li Li, and Zhengbing He. Chatgpt as your vehicle co-pilot: An initial attempt. IEEE Transactions on Intelligent Vehicles, 2023."},{"key":"e_1_3_2_1_20_1","first-page":"23965","volume-title":"Proceedings of the 39th International Conference on Machine Learning (ICML)","author":"Wortsman Mitchell","year":"2022","unstructured":"Mitchell Wortsman, Gabriel Ilharco, Samir Ya Gadre, Rebecca Roelofs, Raphael Gontijo-Lopes, Ari S Morcos, Hongseok Namkoong, Ali Farhadi, Yair Carmon, Simon Kornblith, and Ludwig Schmidt. Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time. In Proceedings of the 39th International Conference on Machine Learning (ICML), pages 23965--23998, 2022."},{"key":"e_1_3_2_1_21_1","volume-title":"PangWei Koh, Chaowei Xiao, and Muhao Chen. Instructional fingerprinting of large language models. CoRR abs\/2401.12255","author":"Xu Jiashu","year":"2024","unstructured":"Jiashu Xu, FeiWang, Mingyu Derek Ma, PangWei Koh, Chaowei Xiao, and Muhao Chen. Instructional fingerprinting of large language models. CoRR abs\/2401.12255, 2024."},{"key":"e_1_3_2_1_22_1","first-page":"7093","volume-title":"Advances in Neural Information Processing Systems (NeurIPS)","volume":"36","author":"Yadav Prateek","year":"2023","unstructured":"Prateek Yadav, Derek Tam, Leshem Choshen, Colin A Raffel, and Mohit Bansal. Ties-merging: Resolving interference when merging models. In A. Oh, T. Neumann, A. Globerson, K. Saenko, M. Hardt, and S. Levine, editors, Advances in Neural Information Processing Systems (NeurIPS), volume 36, pages 7093--7115. Curran Associates, Inc., 2023."},{"key":"e_1_3_2_1_23_1","volume-title":"Language models are super mario: Absorbing abilities from homologous models as a free lunch. CoRR abs\/2311.03099","author":"Yu Le","year":"2024","unstructured":"Le Yu, Bowen Yu, Haiyang Yu, Fei Huang, and Yongbin Li. Language models are super mario: Absorbing abilities from homologous models as a free lunch. CoRR abs\/2311.03099, 2024."},{"key":"e_1_3_2_1_24_1","volume-title":"The Twelfth International Conference on Learning Representations (ICLR)","author":"Zhao Xuandong","year":"2024","unstructured":"Xuandong Zhao, Prabhanjan Vijendra Ananth, Lei Li, and Yu-Xiang Wang. Provable robust watermarking for ai-generated text. In The Twelfth International Conference on Learning Representations (ICLR), 2024."},{"key":"e_1_3_2_1_25_1","volume-title":"Universal and transferable adversarial attacks on aligned language models. CoRR abs\/2307.15043","author":"Zou Andy","year":"2023","unstructured":"Andy Zou, Zifan Wang, Nicholas Carlini, Milad Nasr, J. Zico Kolter, and Matt Fredrikson. Universal and transferable adversarial attacks on aligned language models. CoRR abs\/2307.15043, 2023."}],"event":{"name":"CCS '24: ACM SIGSAC Conference on Computer and Communications Security","location":"Salt Lake City UT USA","acronym":"CCS '24","sponsor":["SIGSAC ACM Special Interest Group on Security, Audit, and Control"]},"container-title":["Proceedings of the 1st ACM Workshop on Large AI Systems and Models with Privacy and Safety Analysis"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3689217.3690614","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3689217.3690614","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,23]],"date-time":"2025-08-23T18:12:55Z","timestamp":1755972775000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3689217.3690614"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,19]]},"references-count":25,"alternative-id":["10.1145\/3689217.3690614","10.1145\/3689217"],"URL":"https:\/\/doi.org\/10.1145\/3689217.3690614","relation":{},"subject":[],"published":{"date-parts":[[2023,11,19]]},"assertion":[{"value":"2024-11-19","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}