{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T06:40:01Z","timestamp":1755844801268,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":91,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T00:00:00Z","timestamp":1733097600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"US Office of Naval Research grant","award":["N00014-23-1-2122"],"award-info":[{"award-number":["N00014-23-1-2122"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,12,2]]},"DOI":"10.1145\/3658644.3670299","type":"proceedings-article","created":{"date-parts":[[2024,12,9]],"date-time":"2024-12-09T12:19:20Z","timestamp":1733746760000},"page":"2281-2295","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Dye4AI: Assuring Data Boundary on Generative AI Services"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7920-7025","authenticated-orcid":false,"given":"Shu","family":"Wang","sequence":"first","affiliation":[{"name":"George Mason University, Fairfax, VA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4152-2107","authenticated-orcid":false,"given":"Kun","family":"Sun","sequence":"additional","affiliation":[{"name":"George Mason University, Fairfax, VA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-4301-9831","authenticated-orcid":false,"given":"Yan","family":"Zhai","sequence":"additional","affiliation":[{"name":"Visa Inc., Ashburn, VA, USA"}]}],"member":"320","published-online":{"date-parts":[[2024,12,9]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3214303"},{"key":"e_1_3_2_1_2_1","unstructured":"Lightning AI. 2023. Lit-GPT. https:\/\/github.com\/Lightning-AI\/lit-gpt."},{"key":"e_1_3_2_1_3_1","unstructured":"Ebtesam Almazrouei Hamza Alobeidli Abdulaziz Alshamsi Alessandro Cappelli Ruxandra Cojocaru Merouane Debbah Etienne Goffinet Daniel Heslow Julien Launay Quentin Malartic Badreddine Noune Baptiste Pannier and Guilherme Penedo. 2023. Falcon-40B: an open large language model with state-of-the-art performance. (2023)."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","unstructured":"Alex Andonian Quentin Anthony Stella Biderman Sid Black Preetham Gali Leo Gao Eric Hallahan Josh Levy-Kramer Connor Leahy Lucas Nestler Kip Parker Michael Pieler Shivanshu Purohit Tri Songz Wang Phil and Samuel Weinbach. 2021. GPT-NeoX: Large Scale Autoregressive Language Modeling in PyTorch. https:\/\/doi.org\/10.5281\/zenodo.5879544","DOI":"10.5281\/zenodo.5879544"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2019.8802997"},{"key":"e_1_3_2_1_6_1","unstructured":"Edward Beeching Cl\u00e9mentine Fourrier Nathan Habib Sheon Han Nathan Lambert Nazneen Rajani Omar Sanseviero Lewis Tunstall and Thomas Wolf. 2023. Open LLM Leaderboard. https:\/\/huggingface.co\/spaces\/HuggingFaceH4\/open_llm_leaderboard."},{"key":"e_1_3_2_1_7_1","volume-title":"ACORN: Input Validation for Secure Aggregation. In 32nd USENIX Security Symposium (USENIX Security 23)","author":"Bell James","year":"2023","unstructured":"James Bell, Adri\u00e0 Gasc\u00f3n, Tancr\u00e8de Lepoint, Baiyu Li, Sarah Meiklejohn, Mariana Raykova, and Cathie Yun. 2023. ACORN: Input Validation for Secure Aggregation. In 32nd USENIX Security Symposium (USENIX Security 23). 4805--4822."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3372297.3417885"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3488659.3493779"},{"key":"e_1_3_2_1_10_1","unstructured":"Manish Bhatt Sahana Chennabasappa Cyrus Nikolaidis Shengye Wan Ivan Evtimov Dominik Gabi Daniel Song Faizan Ahmad Cornelius Aschermann Lorenzo Fontana et al. 2023. Purple Llama CyberSecEval: A Secure Coding Benchmark for Language Models. arXiv preprint arXiv:2312.04724 (2023)."},{"key":"e_1_3_2_1_11_1","unstructured":"Bloomberg. 2023. Microsoft AI Researchers Accidentally Exposed Big Cache of Data. https:\/\/www.bloomberg.com\/news\/articles\/2023-09--18\/microsoft-ai-researchers-accidentally-exposed-big-cache-of-data?embedded-checkout=true [accessed September 2023]."},{"key":"e_1_3_2_1_12_1","unstructured":"Bloomberg. 2023. Samsung Bans Staff's AI Use After Spotting ChatGPT Data Leak. https:\/\/www.bloomberg.com\/news\/articles\/2023-05-02\/samsung-bans-chatgpt-and-other-generative-ai-use-by-staff-after-leak#xj4y7vzkg [accessed July 2023]."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3534642"},{"key":"e_1_3_2_1_14_1","volume-title":"30th USENIX Security Symposium (USENIX Security 21)","author":"Carlini Nicholas","year":"2021","unstructured":"Nicholas Carlini, Florian Tramer, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom Brown, Dawn Song, Ulfar Erlingsson, et al. 2021. Extracting training data from large language models. In 30th USENIX Security Symposium (USENIX Security 21). 2633--2650."},{"key":"e_1_3_2_1_15_1","volume-title":"29th USENIX Security Symposium (USENIX Security 20)","author":"Chandrasekaran Varun","year":"2020","unstructured":"Varun Chandrasekaran, Kamalika Chaudhuri, Irene Giacomelli, Somesh Jha, and Songbai Yan. 2020. Exploring connections between active learning and model extraction. In 29th USENIX Security Symposium (USENIX Security 20). 1309--1326."},{"key":"e_1_3_2_1_16_1","unstructured":"Yupeng Chang Xu Wang Jindong Wang Yuan Wu Linyi Yang Kaijie Zhu Hao Chen Xiaoyuan Yi Cunxiang Wang Yidong Wang et al. 2023. A survey on evaluation of large language models. ACM Transactions on Intelligent Systems and Technology (2023)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3134061"},{"key":"e_1_3_2_1_18_1","volume-title":"Targeted backdoor attacks on deep learning systems using data poisoning. arXiv preprint arXiv:1712.05526","author":"Chen Xinyun","year":"2017","unstructured":"Xinyun Chen, Chang Liu, Bo Li, Kimberly Lu, and Dawn Song. 2017. Targeted backdoor attacks on deep learning systems using data poisoning. arXiv preprint arXiv:1712.05526 (2017)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/SPW50608.2020.00025"},{"key":"e_1_3_2_1_20_1","volume-title":"Fine-tune language models to approximate unbiased in-context learning. arXiv preprint arXiv:2310.03331","author":"Chu Timothy","year":"2023","unstructured":"Timothy Chu, Zhao Song, and Chiwun Yang. 2023. Fine-tune language models to approximate unbiased in-context learning. arXiv preprint arXiv:2310.03331 (2023)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3585385"},{"key":"e_1_3_2_1_22_1","volume-title":"Chatting and cheating: Ensuring academic integrity in the era of ChatGPT. Innovations in Education and Teaching International","author":"Cotton Debby RE","year":"2023","unstructured":"Debby RE Cotton, Peter A Cotton, and J Reuben Shipway. 2023. Chatting and cheating: Ensuring academic integrity in the era of ChatGPT. Innovations in Education and Teaching International (2023), 1--12."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2008.11"},{"key":"e_1_3_2_1_24_1","first-page":"16344","article-title":"Flashattention: Fast and memory-efficient exact attention with io-awareness","volume":"35","author":"Dao Tri","year":"2022","unstructured":"Tri Dao, Dan Fu, Stefano Ermon, Atri Rudra, and Christopher R\u00e9. 2022. Flashattention: Fast and memory-efficient exact attention with io-awareness. Advances in Neural Information Processing Systems, Vol. 35 (2022), 16344--16359.","journal-title":"Advances in Neural Information Processing Systems"},{"volume-title":"28th USENIX security symposium (USENIX security 19). 321--338.","author":"Demontis Ambra","key":"e_1_3_2_1_25_1","unstructured":"Ambra Demontis, Marco Melis, Maura Pintor, Matthew Jagielski, Battista Biggio, Alina Oprea, Cristina Nita-Rotaru, and Fabio Roli. 2019. Why do adversarial attacks transfer? explaining transferability of evasion and poisoning attacks. In 28th USENIX security symposium (USENIX security 19). 321--338."},{"key":"e_1_3_2_1_26_1","volume-title":"Beyond the Safeguards: Exploring the Security Risks of ChatGPT. arXiv preprint arXiv:2305.08005","author":"Derner Erik","year":"2023","unstructured":"Erik Derner and Kristina Batistivc. 2023. Beyond the Safeguards: Exploring the Security Risks of ChatGPT. arXiv preprint arXiv:2305.08005 (2023)."},{"key":"e_1_3_2_1_27_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/SPW53761.2021.00017"},{"key":"e_1_3_2_1_29_1","unstructured":"Peng Gao Jiaming Han Renrui Zhang Ziyi Lin Shijie Geng Aojun Zhou Wei Zhang Pan Lu Conghui He Xiangyu Yue et al. 2023. Llama-adapter v2: Parameter-efficient visual instruction model. arXiv preprint arXiv:2304.15010 (2023)."},{"key":"e_1_3_2_1_30_1","unstructured":"Xinyang Geng and Hao Liu. 2023. OpenLLaMA: An Open Reproduction of LLaMA. https:\/\/github.com\/openlm-research\/open_llama"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3162397"},{"key":"e_1_3_2_1_32_1","volume-title":"an AI experiment by Google. https:\/\/bard.google.com, [accessed","author":"Bard Try","year":"2023","unstructured":"Google. 2023. Try Bard, an AI experiment by Google. https:\/\/bard.google.com, [accessed July 2023]."},{"volume-title":"International Conference on Machine Learning. PMLR, 8056--8071","author":"Guo Chuan","key":"e_1_3_2_1_33_1","unstructured":"Chuan Guo, Brian Karrer, Kamalika Chaudhuri, and Laurens van der Maaten. 2022. Bounding training data reconstruction in private (deep) learning. In International Conference on Machine Learning. PMLR, 8056--8071."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3576915.3623175"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359789.3359824"},{"key":"e_1_3_2_1_36_1","volume-title":"Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685","author":"Hu Edward J","year":"2021","unstructured":"Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2021. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3523273"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2018.00057"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3460120.3485368"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/3400302.3415671"},{"key":"e_1_3_2_1_41_1","volume-title":"30th USENIX Security Symposium (USENIX Security 21)","author":"Jia Hengrui","year":"2021","unstructured":"Hengrui Jia, Christopher A Choquette-Choo, Varun Chandrasekaran, and Nicolas Papernot. 2021. Entangled watermarks as a defense against model extraction. In 30th USENIX Security Symposium (USENIX Security 21). 1937--1954."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/EuroSP.2019.00044"},{"key":"e_1_3_2_1_43_1","volume-title":"Kallista Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, et al.","author":"Kairouz Peter","year":"2021","unstructured":"Peter Kairouz, H Brendan McMahan, Brendan Avent, Aur\u00e9lien Bellet, Mehdi Bennis, Arjun Nitin Bhagoji, Kallista Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, et al. 2021. Advances and open problems in federated learning. Foundations and Trends\u00ae in Machine Learning, Vol. 14, 1--2 (2021), 1--210."},{"key":"e_1_3_2_1_44_1","first-page":"160","article-title":"Abstractive Long Text Summarization using Large Language Models","volume":"12","author":"Keswani Gunjan","year":"2024","unstructured":"Gunjan Keswani, Wani Bisen, Hirkani Padwad, Yash Wankhedkar, Sudhanshu Pandey, and Ayushi Soni. 2024. Abstractive Long Text Summarization using Large Language Models. International Journal of Intelligent Systems and Applications in Engineering, Vol. 12, 12s (2024), 160--168.","journal-title":"International Journal of Intelligent Systems and Applications in Engineering"},{"key":"e_1_3_2_1_45_1","volume-title":"Propile: Probing privacy leakage in large language models. arXiv preprint arXiv:2307.01881","author":"Kim Siwon","year":"2023","unstructured":"Siwon Kim, Sangdoo Yun, Hwaran Lee, Martin Gubri, Sungroh Yoon, and Seong Joon Oh. 2023. Propile: Probing privacy leakage in large language models. arXiv preprint arXiv:2307.01881 (2023)."},{"key":"e_1_3_2_1_46_1","volume-title":"A watermark for large language models. arXiv preprint arXiv:2301.10226","author":"Kirchenbauer John","year":"2023","unstructured":"John Kirchenbauer, Jonas Geiping, Yuxin Wen, Jonathan Katz, Ian Miers, and Tom Goldstein. 2023. A watermark for large language models. arXiv preprint arXiv:2301.10226 (2023)."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.397"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3615017"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i12.26750"},{"key":"e_1_3_2_1_50_1","volume-title":"Privacy-preserving prompt tuning for large language model services. arXiv preprint arXiv:2305.06212","author":"Li Yansong","year":"2023","unstructured":"Yansong Li, Zhixing Tan, and Yang Liu. 2023. Privacy-preserving prompt tuning for large language model services. arXiv preprint arXiv:2305.06212 (2023)."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3372297.3423362"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2021.3108434"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3319535.3363216"},{"key":"e_1_3_2_1_54_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_1_55_1","volume-title":"Analyzing leakage of personally identifiable information in language models. arXiv preprint arXiv:2302.00539","author":"Lukas Nils","year":"2023","unstructured":"Nils Lukas, Ahmed Salem, Robert Sim, Shruti Tople, Lukas Wutschitz, and Santiago Zanella-B\u00e9guelin. 2023. Analyzing leakage of personally identifiable information in language models. arXiv preprint arXiv:2302.00539 (2023)."},{"volume-title":"Presidio: Data Protection and De-identification SDK. https:\/\/microsoft.github.io\/presidio\/, [accessed","year":"2023","key":"e_1_3_2_1_56_1","unstructured":"Microsoft. 2023. Presidio: Data Protection and De-identification SDK. https:\/\/microsoft.github.io\/presidio\/, [accessed July 2023]."},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.5555\/1858842.1858883"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.future.2020.10.007"},{"key":"e_1_3_2_1_59_1","volume-title":"API data usage policies. https:\/\/openai.com\/policies\/api-data-usage-policies, [accessed","author":"AI.","year":"2023","unstructured":"OpenAI. 2023. API data usage policies. https:\/\/openai.com\/policies\/api-data-usage-policies, [accessed July 2023]."},{"key":"e_1_3_2_1_60_1","volume-title":"Data usage for consumer services FAQ. https:\/\/help.openai.com\/en\/articles\/7039943-data-usage-for-consumer-services-faq, [accessed","author":"AI.","year":"2023","unstructured":"OpenAI. 2023. Data usage for consumer services FAQ. https:\/\/help.openai.com\/en\/articles\/7039943-data-usage-for-consumer-services-faq, [accessed July 2023]."},{"key":"e_1_3_2_1_61_1","volume-title":"https:\/\/openai.com\/blog\/chatgpt, [accessed","author":"Introducing AI.","year":"2023","unstructured":"OpenAI. 2023. Introducing ChatGPT. https:\/\/openai.com\/blog\/chatgpt, [accessed July 2023]."},{"key":"e_1_3_2_1_62_1","volume-title":"Enterprise privacy at OpenAI. https:\/\/openai.com\/enterprise-privacy, [accessed","author":"AI.","year":"2024","unstructured":"OpenAI. 2024. Enterprise privacy at OpenAI. https:\/\/openai.com\/enterprise-privacy, [accessed Jan 2024]."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/3052973.3053009"},{"key":"e_1_3_2_1_64_1","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans Ilya Sutskever et al. 2018. Improving language understanding by generative pre-training. (2018)."},{"key":"e_1_3_2_1_65_1","volume-title":"Art-ificial Intelligence: The Effect of AI Disclosure on Evaluations of Creative Content. arXiv preprint arXiv:2303.06217","author":"Raj Manav","year":"2023","unstructured":"Manav Raj, Justin Berg, and Rob Seamans. 2023. Art-ificial Intelligence: The Effect of AI Disclosure on Evaluations of Creative Content. arXiv preprint arXiv:2303.06217 (2023)."},{"key":"e_1_3_2_1_66_1","volume-title":"Warm: On the benefits of weight averaged reward models. arXiv preprint arXiv:2401.12187","author":"Ram\u00e9 Alexandre","year":"2024","unstructured":"Alexandre Ram\u00e9, Nino Vieillard, L\u00e9onard Hussenot, Robert Dadashi, Geoffrey Cideron, Olivier Bachem, and Johan Ferret. 2024. Warm: On the benefits of weight averaged reward models. arXiv preprint arXiv:2401.12187 (2024)."},{"volume-title":"29th USENIX security symposium (USENIX Security 20). 1291--1308.","author":"Salem Ahmed","key":"e_1_3_2_1_67_1","unstructured":"Ahmed Salem, Apratim Bhattacharya, Michael Backes, Mario Fritz, and Yang Zhang. 2020. Updates-Leak: Data set inference and reconstruction attacks in online learning. In 29th USENIX security symposium (USENIX Security 20). 1291--1308."},{"key":"e_1_3_2_1_68_1","volume-title":"Lost at c: A user study on the security implications of large language model code assistants. arXiv preprint arXiv:2208.09727","author":"Sandoval Gustavo","year":"2023","unstructured":"Gustavo Sandoval, Hammond Pearce, Teo Nys, Ramesh Karri, Siddharth Garg, and Brendan Dolan-Gavitt. 2023. Lost at c: A user study on the security implications of large language model code assistants. arXiv preprint arXiv:2208.09727 (2023)."},{"key":"e_1_3_2_1_69_1","volume-title":"Clustered federated learning: Model-agnostic distributed multitask optimization under privacy constraints","author":"Sattler Felix","year":"2020","unstructured":"Felix Sattler, Klaus-Robert M\u00fcller, and Wojciech Samek. 2020. Clustered federated learning: Model-agnostic distributed multitask optimization under privacy constraints. IEEE transactions on neural networks and learning systems, Vol. 32, 8 (2020), 3710--3722."},{"volume-title":"30th USENIX security symposium (USENIX security 21). 1487--1504.","author":"Severi Giorgio","key":"e_1_3_2_1_70_1","unstructured":"Giorgio Severi, Jim Meyer, Scott Coull, and Alina Oprea. 2021. Explanation-Guided backdoor poisoning attacks against malware classifiers. In 30th USENIX security symposium (USENIX security 21). 1487--1504."},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1155\/2021\/6641832"},{"key":"e_1_3_2_1_72_1","volume-title":"USENIX Sec. Symp. USENIX Association","volume":"8","author":"Shan Shawn","year":"2022","unstructured":"Shawn Shan, Arjun Nitin Bhagoji, Haitao Zheng, and Ben Y Zhao. 2022. Traceback of targeted data poisoning attacks in neural networks. In USENIX Sec. Symp. USENIX Association, Vol. 8."},{"key":"e_1_3_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP46214.2022.9833647"},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1109\/STARTUP.2016.7583912"},{"key":"e_1_3_2_1_75_1","volume-title":"31st USENIX Security Symposium (USENIX Security 22)","author":"Stevens Timothy","year":"2022","unstructured":"Timothy Stevens, Christian Skalka, Christelle Vincent, John Ring, Samuel Clark, and Joseph Near. 2022. Efficient differentially private secure aggregation for federated learning via hardness of learning with errors. In 31st USENIX Security Symposium (USENIX Security 22). 1379--1395."},{"key":"e_1_3_2_1_76_1","volume-title":"27th USENIX Security Symposium (USENIX Security 18)","author":"Suciu Octavian","year":"2018","unstructured":"Octavian Suciu, Radu Marginean, Yigitcan Kaya, Hal Daume III, and Tudor Dumitras. 2018. When does machine learning FAIL? generalized transferability for evasion and poisoning attacks. In 27th USENIX Security Symposium (USENIX Security 18). 1299--1316."},{"key":"e_1_3_2_1_77_1","volume-title":"Hashimoto","author":"Taori Rohan","year":"2023","unstructured":"Rohan Taori, Ishaan Gulrajani, Tianyi Zhang, Yann Dubois, Xuechen Li, Carlos Guestrin, Percy Liang, and Tatsunori B. Hashimoto. 2023. Stanford Alpaca: An Instruction-following LLaMA model. https:\/\/github.com\/tatsu-lab\/stanford_alpaca."},{"key":"e_1_3_2_1_78_1","unstructured":"Paul Vanhaesebrouck Aur\u00e9lien Bellet and Marc Tommasi. 2017. Decentralized collaborative learning of personalized models over networks. In Artificial Intelligence and Statistics. PMLR 509--517."},{"key":"e_1_3_2_1_79_1","unstructured":"Kushala VM Harikrishna Warrier Yogesh Gupta et al. 2024. Fine Tuning LLM for Enterprise: Practical Guidelines and Recommendations. arXiv preprint arXiv:2404.10779 (2024)."},{"key":"e_1_3_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2019.00031"},{"key":"e_1_3_2_1_81_1","volume-title":"International Conference on Machine Learning. PMLR, 22769--22783","author":"Wang Wenxiao","year":"2022","unstructured":"Wenxiao Wang, Alexander J Levine, and Soheil Feizi. 2022. Improved certified defenses against data poisoning with (deterministic) finite aggregation. In International Conference on Machine Learning. PMLR, 22769--22783."},{"key":"e_1_3_2_1_82_1","volume-title":"A prompt pattern catalog to enhance prompt engineering with chatgpt. arXiv preprint arXiv:2302.11382","author":"White Jules","year":"2023","unstructured":"Jules White, Quchen Fu, Sam Hays, Michael Sandborn, Carlos Olea, Henry Gilbert, Ashraf Elnashar, Jesse Spencer-Smith, and Douglas C Schmidt. 2023. A prompt pattern catalog to enhance prompt engineering with chatgpt. arXiv preprint arXiv:2302.11382 (2023)."},{"key":"e_1_3_2_1_83_1","unstructured":"Alan Wong Vincent Lacey Chaitanya Gharpure Rebecca Hao Priya Venkatraman Gal Elidan Roee Engelberg Lidan Hackmon Roni Rabin Michael Fink et al. 2023. Reading Comprehension Assessment Using LLM-based Chatbot. (2023)."},{"key":"e_1_3_2_1_84_1","volume-title":"30th USENIX Security Symposium (USENIX Security 21)","author":"Xi Zhaohan","year":"2021","unstructured":"Zhaohan Xi, Ren Pang, Shouling Ji, and Ting Wang. 2021. Graph backdoor. In 30th USENIX Security Symposium (USENIX Security 21). 1523--1540."},{"key":"e_1_3_2_1_85_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP40001.2021.00034"},{"key":"e_1_3_2_1_86_1","doi-asserted-by":"publisher","DOI":"10.1145\/3319535.3354209"},{"key":"e_1_3_2_1_87_1","volume-title":"Balancing Speciality and Versatility: a Coarse to Fine Framework for Supervised Fine-tuning Large Language Model. arXiv preprint arXiv:2404.10306","author":"Zhang Hengyuan","year":"2024","unstructured":"Hengyuan Zhang, Yanru Wu, Dawei Li, Zacc Yang, Rui Zhao, Yong Jiang, and Fei Tan. 2024. Balancing Speciality and Versatility: a Coarse to Fine Framework for Supervised Fine-tuning Large Language Model. arXiv preprint arXiv:2404.10306 (2024)."},{"key":"e_1_3_2_1_88_1","volume-title":"Llama-adapter: Efficient fine-tuning of language models with zero-init attention. arXiv preprint arXiv:2303.16199","author":"Zhang Renrui","year":"2023","unstructured":"Renrui Zhang, Jiaming Han, Aojun Zhou, Xiangfei Hu, Shilin Yan, Pan Lu, Hongsheng Li, Peng Gao, and Yu Qiao. 2023. Llama-adapter: Efficient fine-tuning of language models with zero-init attention. arXiv preprint arXiv:2303.16199 (2023)."},{"key":"e_1_3_2_1_89_1","unstructured":"Shengyu Zhang Linfeng Dong Xiaoya Li Sen Zhang Xiaofei Sun Shuhe Wang Jiwei Li Runyi Hu Tianwei Zhang Fei Wu et al. 2023. Instruction tuning for large language models: A survey. arXiv preprint arXiv:2308.10792 (2023)."},{"key":"e_1_3_2_1_90_1","volume-title":"31st USENIX Security Symposium (USENIX Security 22)","author":"Zhang Zhikun","year":"2022","unstructured":"Zhikun Zhang, Min Chen, Michael Backes, Yun Shen, and Yang Zhang. 2022. Inference attacks against graph neural networks. In 31st USENIX Security Symposium (USENIX Security 22). 4543--4560."},{"key":"e_1_3_2_1_91_1","doi-asserted-by":"publisher","DOI":"10.1145\/3374664.3375751"}],"event":{"name":"CCS '24: ACM SIGSAC Conference on Computer and Communications Security","sponsor":["SIGSAC ACM Special Interest Group on Security, Audit, and Control"],"location":"Salt Lake City UT USA","acronym":"CCS '24"},"container-title":["Proceedings of the 2024 on ACM SIGSAC Conference on Computer and Communications Security"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3658644.3670299","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3658644.3670299","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T06:03:47Z","timestamp":1755842627000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3658644.3670299"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,2]]},"references-count":91,"alternative-id":["10.1145\/3658644.3670299","10.1145\/3658644"],"URL":"https:\/\/doi.org\/10.1145\/3658644.3670299","relation":{},"subject":[],"published":{"date-parts":[[2024,12,2]]},"assertion":[{"value":"2024-12-09","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}