{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T18:49:43Z","timestamp":1776106183547,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":65,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,8,8]],"date-time":"2023-08-08T00:00:00Z","timestamp":1691452800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,8,8]]},"DOI":"10.1145\/3600211.3604690","type":"proceedings-article","created":{"date-parts":[[2023,8,29]],"date-time":"2023-08-29T18:41:37Z","timestamp":1693334497000},"page":"287-296","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":11,"title":["Self-Destructing Models: Increasing the Costs of Harmful Dual Uses of Foundation Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3938-0541","authenticated-orcid":false,"given":"Peter","family":"Henderson","sequence":"first","affiliation":[{"name":"Stanford University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7487-1744","authenticated-orcid":false,"given":"Eric","family":"Mitchell","sequence":"additional","affiliation":[{"name":"Stanford University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6155-649X","authenticated-orcid":false,"given":"Christopher","family":"Manning","sequence":"additional","affiliation":[{"name":"Stanford University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6459-7745","authenticated-orcid":false,"given":"Dan","family":"Jurafsky","sequence":"additional","affiliation":[{"name":"Stanford University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6298-0874","authenticated-orcid":false,"given":"Chelsea","family":"Finn","sequence":"additional","affiliation":[{"name":"Stanford University, USA"}]}],"member":"320","published-online":{"date-parts":[[2023,8,29]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"A. Agrawal B. Amos S. Barratt S. Boyd S. Diamond and Z. Kolter. 2019. Differentiable Convex Optimization Layers. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_1_2_1","volume-title":"Fine tuning CLIP with Remote Sensing (Satellite) images and captions. HuggingFace Blog","author":"Arutiunian Artashes","year":"2021","unstructured":"Artashes Arutiunian, Dev Vidhani, Goutham Venkatesh, Mayank Bhaskar, Ritobrata Ghosh, and Sujit Pal. 2021. Fine tuning CLIP with Remote Sensing (Satellite) images and captions. HuggingFace Blog (2021). https:\/\/huggingface.co\/blog\/fine-tune-clip-rsicd"},{"key":"e_1_3_2_1_3_1","volume-title":"Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862","author":"Bai Yuntao","year":"2022","unstructured":"Yuntao Bai, Andy Jones, Kamal Ndousse, Amanda Askell, Anna Chen, Nova DasSarma, Dawn Drain, Stanislav Fort, Deep Ganguli, Tom Henighan, 2022. Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862 (2022)."},{"key":"e_1_3_2_1_4_1","volume-title":"Algorithms for hyper-parameter optimization. Advances in neural information processing systems 24","author":"Bergstra James","year":"2011","unstructured":"James Bergstra, R\u00e9mi Bardenet, Yoshua Bengio, and Bal\u00e1zs K\u00e9gl. 2011. Algorithms for hyper-parameter optimization. Advances in neural information processing systems 24 (2011)."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.25080\/Majora-8b375195-003"},{"key":"e_1_3_2_1_6_1","volume-title":"GPT-NeoX-20B: An Open-Source Autoregressive Language Model. arXiv preprint arXiv:2204.06745","author":"Black Sid","year":"2022","unstructured":"Sid Black, Stella Biderman, Eric Hallahan, Quentin Anthony, Leo Gao, Laurence Golding, Horace He, Connor Leahy, Kyle McDonell, Jason Phang, 2022. GPT-NeoX-20B: An Open-Source Autoregressive Language Model. arXiv preprint arXiv:2204.06745 (2022)."},{"key":"e_1_3_2_1_7_1","volume-title":"On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258","author":"Bommasani Rishi","year":"2021","unstructured":"Rishi Bommasani, Drew\u00a0A Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael\u00a0S Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, 2021. On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258 (2021)."},{"key":"e_1_3_2_1_8_1","volume-title":"The malicious use of artificial intelligence: Forecasting, prevention, and mitigation. arXiv preprint arXiv:1802.07228","author":"Brundage Miles","year":"2018","unstructured":"Miles Brundage, Shahar Avin, Jack Clark, Helen Toner, Peter Eckersley, Ben Garfinkel, Allan Dafoe, Paul Scharre, Thomas Zeitzoff, Bobby Filar, 2018. The malicious use of artificial intelligence: Forecasting, prevention, and mitigation. arXiv preprint arXiv:1802.07228 (2018)."},{"key":"e_1_3_2_1_9_1","volume-title":"Toward trustworthy AI development: mechanisms for supporting verifiable claims. arXiv preprint arXiv:2004.07213","author":"Brundage Miles","year":"2020","unstructured":"Miles Brundage, Shahar Avin, Jasmine Wang, Haydn Belfield, Gretchen Krueger, Gillian Hadfield, Heidy Khlaaf, Jingying Yang, Helen Toner, Ruth Fong, 2020. Toward trustworthy AI development: mechanisms for supporting verifiable claims. arXiv preprint arXiv:2004.07213 (2020)."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287572"},{"key":"e_1_3_2_1_11_1","volume-title":"Editing Factual Knowledge in Language Models. ArXiv abs\/2104.08164","author":"Cao Nicola De","year":"2021","unstructured":"Nicola De Cao, W. Aziz, and Ivan Titov. 2021. Editing Factual Knowledge in Language Models. ArXiv abs\/2104.08164 (2021)."},{"key":"e_1_3_2_1_12_1","volume-title":"8-bit Matrix Multiplication for Transformers at Scale. arXiv preprint arXiv:2208.07339","author":"Dettmers Tim","year":"2022","unstructured":"Tim Dettmers, Mike Lewis, Younes Belkada, and Luke Zettlemoyer. 2022. LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale. arXiv preprint arXiv:2208.07339 (2022)."},{"key":"e_1_3_2_1_13_1","volume-title":"The case for 4-bit precision: k-bit Inference Scaling Laws. arXiv preprint arXiv:2212.09720","author":"Dettmers Tim","year":"2022","unstructured":"Tim Dettmers and Luke Zettlemoyer. 2022. The case for 4-bit precision: k-bit Inference Scaling Laws. arXiv preprint arXiv:2212.09720 (2022)."},{"key":"e_1_3_2_1_14_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.5555\/2946645.3007036"},{"key":"e_1_3_2_1_16_1","volume-title":"Censoring representations with an adversary. arXiv preprint arXiv:1511.05897","author":"Edwards Harrison","year":"2015","unstructured":"Harrison Edwards and Amos Storkey. 2015. Censoring representations with an adversary. arXiv preprint arXiv:1511.05897 (2015)."},{"key":"e_1_3_2_1_17_1","volume-title":"The EU\u2019s attempt to regulate open-source AI is counterproductive. Brookings TechTank","author":"Engler Alex","year":"2022","unstructured":"Alex Engler. 2022. The EU\u2019s attempt to regulate open-source AI is counterproductive. Brookings TechTank (2022)."},{"key":"e_1_3_2_1_18_1","unstructured":"Carlos\u00a0Mu\u00f1oz Ferrandis. 2022. OpenRAIL: Towards open and responsible AI licensing frameworks. https:\/\/huggingface.co\/blog\/open_rail."},{"key":"e_1_3_2_1_19_1","volume-title":"Proceedings of the 34th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a070)","author":"Finn Chelsea","year":"2017","unstructured":"Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. In Proceedings of the 34th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a070), Doina Precup and Yee\u00a0Whye Teh (Eds.). PMLR, 1126\u20131135. https:\/\/proceedings.mlr.press\/v70\/finn17a.html"},{"key":"e_1_3_2_1_20_1","volume-title":"Meta-Learning with Warped Gradient Descent. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=rkeiQlBFPB","author":"Flennerhag Sebastian","year":"2020","unstructured":"Sebastian Flennerhag, Andrei\u00a0A. Rusu, Razvan Pascanu, Francesco Visin, Hujun Yin, and Raia Hadsell. 2020. Meta-Learning with Warped Gradient Descent. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=rkeiQlBFPB"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.51593\/20190001"},{"key":"e_1_3_2_1_22_1","volume-title":"International conference on machine learning. PMLR, 1180\u20131189","author":"Ganin Yaroslav","year":"2015","unstructured":"Yaroslav Ganin and Victor Lempitsky. 2015. Unsupervised domain adaptation by backpropagation. In International conference on machine learning. PMLR, 1180\u20131189."},{"key":"e_1_3_2_1_23_1","volume-title":"Hackers are selling a service that bypasses ChatGPT restrictions on malware. arstechnica","author":"Goodin Dan","year":"2023","unstructured":"Dan Goodin. 2023. Hackers are selling a service that bypasses ChatGPT restrictions on malware. arstechnica (2023)."},{"key":"e_1_3_2_1_24_1","volume-title":"Generalized Inner Loop Meta-Learning. arXiv preprint arXiv:1910.01727","author":"Grefenstette Edward","year":"2019","unstructured":"Edward Grefenstette, Brandon Amos, Denis Yarats, Phu\u00a0Mon Htut, Artem Molchanov, Franziska Meier, Douwe Kiela, Kyunghyun Cho, and Soumith Chintala. 2019. Generalized Inner Loop Meta-Learning. arXiv preprint arXiv:1910.01727 (2019)."},{"key":"e_1_3_2_1_25_1","volume-title":"Regulating ChatGPT and other Large Generative AI Models. arXiv preprint arXiv:2302.02337","author":"Hacker Philipp","year":"2023","unstructured":"Philipp Hacker, Andreas Engel, and Marco Mauer. 2023. Regulating ChatGPT and other Large Generative AI Models. arXiv preprint arXiv:2302.02337 (2023)."},{"key":"e_1_3_2_1_26_1","volume-title":"Proceedings of the 36th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a097)","author":"Houlsby Neil","year":"2019","unstructured":"Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin De\u00a0Laroussilhe, Andrea Gesmundo, Mona Attariyan, and Sylvain Gelly. 2019. Parameter-Efficient Transfer Learning for NLP. In Proceedings of the 36th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a097), Kamalika Chaudhuri and Ruslan Salakhutdinov (Eds.). PMLR, 2790\u20132799. https:\/\/proceedings.mlr.press\/v97\/houlsby19a.html"},{"key":"e_1_3_2_1_27_1","unstructured":"Edward Hu Yelong Shen Phil Wallis Zeyuan Allen-Zhu Yuanzhi Li Lu Wang and Weizhu Chen. 2021. LoRA: Low-Rank Adaptation of Large Language Models. arxiv:2106.09685\u00a0[cs.CL]"},{"key":"e_1_3_2_1_28_1","volume-title":"AI porn is easy to make now. For women, that\u2019s a nightmare.The Washington Post","author":"Hunter Tatum","year":"2023","unstructured":"Tatum Hunter. 2023. AI porn is easy to make now. For women, that\u2019s a nightmare.The Washington Post (2023)."},{"key":"e_1_3_2_1_29_1","volume-title":"Pretraining Language Models with Human Preferences. arXiv preprint arXiv:2302.08582","author":"Korbak Tomasz","year":"2023","unstructured":"Tomasz Korbak, Kejian Shi, Angelica Chen, Rasika Bhalerao, Christopher\u00a0L Buckley, Jason Phang, Samuel\u00a0R Bowman, and Ethan Perez. 2023. Pretraining Language Models with Human Preferences. arXiv preprint arXiv:2302.08582 (2023)."},{"key":"e_1_3_2_1_30_1","volume-title":"International Conference on Machine Learning. 2933\u20132942","author":"Lee Yoonho","year":"2018","unstructured":"Yoonho Lee and Seungjin Choi. 2018. Gradient-based meta-learning with learned layerwise metric and subspace. In International Conference on Machine Learning. 2933\u20132942."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00566"},{"key":"e_1_3_2_1_32_1","unstructured":"Xiang\u00a0Lisa Li and Percy Liang. 2021. Prefix-Tuning: Optimizing Continuous Prompts for Generation. arxiv:2101.00190\u00a0[cs.CL]"},{"key":"e_1_3_2_1_33_1","volume-title":"Meta-SGD: Learning to Learn Quickly for Few Shot Learning. CoRR abs\/1707.09835","author":"Li Zhenguo","year":"2017","unstructured":"Zhenguo Li, Fengwei Zhou, Fei Chen, and Hang Li. 2017. Meta-SGD: Learning to Learn Quickly for Few Shot Learning. CoRR abs\/1707.09835 (2017). arXiv:1707.09835http:\/\/arxiv.org\/abs\/1707.09835"},{"key":"e_1_3_2_1_34_1","unstructured":"Percy Liang Rishi Bommasani Kathleen\u00a0A. Creel and Rob Reich. 2022. The Time Is Now to Develop Community Norms for the Release of Foundation Models. https:\/\/crfm.stanford.edu\/2022\/05\/17\/community-norms.html"},{"key":"e_1_3_2_1_35_1","volume-title":"How Would The Viewer Feel? Estimating Wellbeing From Video Scenarios. arXiv preprint arXiv:2210.10039","author":"Mazeika Mantas","year":"2022","unstructured":"Mantas Mazeika, Eric Tang, Andy Zou, Steven Basart, Jun\u00a0Shern Chan, Dawn Song, David Forsyth, Jacob Steinhardt, and Dan Hendrycks. 2022. How Would The Viewer Feel? Estimating Wellbeing From Video Scenarios. arXiv preprint arXiv:2210.10039 (2022)."},{"key":"e_1_3_2_1_36_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=0DcZxeWfOPt","author":"Mitchell Eric","year":"2022","unstructured":"Eric Mitchell, Charles Lin, Antoine Bosselut, Chelsea Finn, and Christopher\u00a0D Manning. 2022. Fast Model Editing at Scale. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=0DcZxeWfOPt"},{"key":"e_1_3_2_1_37_1","volume-title":"Memory-Based Model Editing at Scale. arXiv preprint arXiv:2206.06520","author":"Mitchell Eric","year":"2022","unstructured":"Eric Mitchell, Charles Lin, Antoine Bosselut, Christopher\u00a0D Manning, and Chelsea Finn. 2022. Memory-Based Model Editing at Scale. arXiv preprint arXiv:2206.06520 (2022)."},{"key":"e_1_3_2_1_38_1","volume-title":"Ethics of Data and Analytics","author":"Olson Parmy","unstructured":"Parmy Olson. 2022. The Quiet Growth of Race-Detection Software Sparks Concerns over Bias. In Ethics of Data and Analytics. Auerbach Publications, 201\u2013205."},{"key":"e_1_3_2_1_39_1","unstructured":"Laurent Orseau and MS Armstrong. 2016. Safely interruptible agents. (2016)."},{"key":"e_1_3_2_1_40_1","volume-title":"Training language models to follow instructions with human feedback. arXiv preprint arXiv:2203.02155","author":"Ouyang Long","year":"2022","unstructured":"Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll\u00a0L Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, 2022. Training language models to follow instructions with human feedback. arXiv preprint arXiv:2203.02155 (2022)."},{"key":"e_1_3_2_1_41_1","volume-title":"Reducing malicious use of synthetic media research: Considerations and potential release practices for machine learning. arXiv preprint arXiv:1907.11274","author":"Ovadya Aviv","year":"2019","unstructured":"Aviv Ovadya and Jess Whittlestone. 2019. Reducing malicious use of synthetic media research: Considerations and potential release practices for machine learning. arXiv preprint arXiv:1907.11274 (2019)."},{"key":"e_1_3_2_1_42_1","volume-title":"Advances in Neural Information Processing Systems, H.\u00a0Wallach, H.\u00a0Larochelle, A.\u00a0Beygelzimer, F.\u00a0d'Alch\u00e9-Buc, E.\u00a0Fox, and R.\u00a0Garnett (Eds.). Vol.\u00a032. Curran Associates","author":"Park Eunbyung","unstructured":"Eunbyung Park and Junier\u00a0B Oliva. 2019. Meta-Curvature. In Advances in Neural Information Processing Systems, H.\u00a0Wallach, H.\u00a0Larochelle, A.\u00a0Beygelzimer, F.\u00a0d'Alch\u00e9-Buc, E.\u00a0Fox, and R.\u00a0Garnett (Eds.). Vol.\u00a032. Curran Associates, Inc."},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1146"},{"key":"e_1_3_2_1_44_1","volume-title":"International conference on machine learning. PMLR, 8748\u20138763","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong\u00a0Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, 2021. Learning transferable visual models from natural language supervision. In International conference on machine learning. PMLR, 8748\u20138763."},{"key":"e_1_3_2_1_45_1","volume-title":"Red-Teaming the Stable Diffusion Safety Filter. arXiv preprint arXiv:2210.04610","author":"Rando Javier","year":"2022","unstructured":"Javier Rando, Daniel Paleka, David Lindner, Lennard Heim, and Florian Tram\u00e8r. 2022. Red-Teaming the Stable Diffusion Safety Filter. arXiv preprint arXiv:2210.04610 (2022)."},{"key":"e_1_3_2_1_46_1","volume-title":"Linear Adversarial Concept Erasure. arXiv preprint arXiv:2201.12091","author":"Ravfogel Shauli","year":"2022","unstructured":"Shauli Ravfogel, Michael Twiton, Yoav Goldberg, and Ryan Cotterell. 2022. Linear Adversarial Concept Erasure. arXiv preprint arXiv:2201.12091 (2022)."},{"key":"e_1_3_2_1_47_1","volume-title":"Adversarial Concept Erasure in Kernel Space. arXiv preprint arXiv:2201.12191","author":"Ravfogel Shauli","year":"2022","unstructured":"Shauli Ravfogel, Francisco Vargas, Yoav Goldberg, and Ryan Cotterell. 2022. Adversarial Concept Erasure in Kernel Space. arXiv preprint arXiv:2201.12191 (2022)."},{"key":"e_1_3_2_1_48_1","unstructured":"Robin Rombach Andreas Blattmann Dominik Lorenz Patrick Esser and Bj\u00f6rn Ommer. 2021. High-Resolution Image Synthesis with Latent Diffusion Models. arxiv:2112.10752\u00a0[cs.CV]"},{"key":"e_1_3_2_1_49_1","volume-title":"Structured access to AI capabilities: an emerging paradigm for safe AI deployment. arXiv preprint arXiv:2201.05159","author":"Shevlane Toby","year":"2022","unstructured":"Toby Shevlane. 2022. Structured access to AI capabilities: an emerging paradigm for safe AI deployment. arXiv preprint arXiv:2201.05159 (2022)."},{"key":"e_1_3_2_1_50_1","volume-title":"Editable Neural Networks. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=HJedXaEtvS","author":"Sinitsin Anton","year":"2020","unstructured":"Anton Sinitsin, Vsevolod Plokhotnyuk, Dmitry Pyrkin, Sergei Popov, and Artem Babenko. 2020. Editable Neural Networks. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=HJedXaEtvS"},{"key":"e_1_3_2_1_51_1","volume-title":"The Gradient of Generative AI Release: Methods and Considerations. arXiv preprint arXiv:2302.04844","author":"Solaiman Irene","year":"2023","unstructured":"Irene Solaiman. 2023. The Gradient of Generative AI Release: Methods and Considerations. arXiv preprint arXiv:2302.04844 (2023)."},{"key":"e_1_3_2_1_52_1","volume-title":"Release strategies and the social impacts of language models. arXiv preprint arXiv:1908.09203","author":"Solaiman Irene","year":"2019","unstructured":"Irene Solaiman, Miles Brundage, Jack Clark, Amanda Askell, Ariel Herbert-Voss, Jeff Wu, Alec Radford, Gretchen Krueger, Jong\u00a0Wook Kim, Sarah Kreps, 2019. Release strategies and the social impacts of language models. arXiv preprint arXiv:1908.09203 (2019)."},{"key":"e_1_3_2_1_53_1","volume-title":"LLaMA: Open and Efficient Foundation Language Models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, 2023. LLaMA: Open and Efficient Foundation Language Models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_1_54_1","unstructured":"Iulia Turc Ming-Wei Chang Kenton Lee and Kristina Toutanova. 2020. Well-Read Students Learn Better: On the Importance of Pre-training Compact Models. https:\/\/openreview.net\/forum?id=BJg7x1HFvB"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-022-00465-9"},{"key":"e_1_3_2_1_56_1","first-page":"62186","article-title":"Implementation of Additional Export Controls: Certain Advanced Computing and Semiconductor Manufacturing Items; Supercomputer and Semiconductor End Use; Entity List Modification","volume":"87","author":"U.S. Department of Commerce.","year":"2022","unstructured":"U.S. Department of Commerce. 2022. Implementation of Additional Export Controls: Certain Advanced Computing and Semiconductor Manufacturing Items; Supercomputer and Semiconductor End Use; Entity List Modification. Federal Register 87 (2022), 62186. https:\/\/www.federalregister.gov\/documents\/2022\/10\/13\/2022-21658\/implementation-of-additional-export-controls-certain-advanced-computing-and-semiconductor","journal-title":"Federal Register"},{"key":"e_1_3_2_1_57_1","volume-title":"YouTuber trains AI bot on 4chan\u2019s pile o\u2019bile with entirely predictable results. The Verge","author":"Vincent James","year":"2022","unstructured":"James Vincent. 2022. YouTuber trains AI bot on 4chan\u2019s pile o\u2019bile with entirely predictable results. The Verge (2022)."},{"key":"e_1_3_2_1_58_1","volume-title":"Meta\u2019s powerful AI language model has leaked online \u2014 what happens now?The Verge","author":"Vincent James","year":"2023","unstructured":"James Vincent. 2023. Meta\u2019s powerful AI language model has leaked online \u2014 what happens now?The Verge (2023)."},{"key":"e_1_3_2_1_59_1","volume-title":"The tension between openness and prudence in AI research. arXiv preprint arXiv:1910.01170","author":"Whittlestone Jess","year":"2019","unstructured":"Jess Whittlestone and Aviv Ovadya. 2019. The tension between openness and prudence in AI research. arXiv preprint arXiv:1910.01170 (2019)."},{"key":"e_1_3_2_1_60_1","volume-title":"Improving out-of-distribution robustness via selective augmentation. arXiv preprint arXiv:2201.00299","author":"Yao Huaxiu","year":"2022","unstructured":"Huaxiu Yao, Yu Wang, Sai Li, Linjun Zhang, Weixin Liang, James Zou, and Chelsea Finn. 2022. Improving out-of-distribution robustness via selective augmentation. arXiv preprint arXiv:2201.00299 (2022)."},{"key":"e_1_3_2_1_61_1","volume-title":"Democratizing access to large-scale language models with OPT-175B. Meta AI","author":"Zhang Susan","year":"2022","unstructured":"Susan Zhang, Mona Diab, and Luke Zettlemoyer. 2022. Democratizing access to large-scale language models with OPT-175B. Meta AI (2022)."},{"key":"e_1_3_2_1_62_1","volume-title":"OPT: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068","author":"Zhang Susan","year":"2022","unstructured":"Susan Zhang, Stephen Roller, Naman Goyal, Mikel Artetxe, Moya Chen, Shuohui Chen, Christopher Dewan, Mona Diab, Xian Li, Xi\u00a0Victoria Lin, 2022. OPT: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068 (2022)."},{"key":"e_1_3_2_1_63_1","volume-title":"Domain generalization with optimal transport and metric learning. arXiv preprint arXiv:2007.10573","author":"Zhou Fan","year":"2020","unstructured":"Fan Zhou, Zhuqing Jiang, Changjian Shui, Boyu Wang, and Brahim Chaib-draa. 2020. Domain generalization with optimal transport and metric learning. arXiv preprint arXiv:2007.10573 (2020)."},{"key":"e_1_3_2_1_64_1","volume-title":"Thirty-sixth International Conference on Machine Learning (ICML 2019)","author":"Zintgraf M","year":"2019","unstructured":"Luisa\u00a0M Zintgraf, Kyriacos Shiarlis, Vitaly Kurin, Katja Hofmann, and Shimon Whiteson. 2019. Fast Context Adaptation via Meta-Learning. Thirty-sixth International Conference on Machine Learning (ICML 2019) (2019)."},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.51593\/20190007"}],"event":{"name":"AIES '23: AAAI\/ACM Conference on AI, Ethics, and Society","location":"Montr\u00e9al QC Canada","acronym":"AIES '23","sponsor":["SIGAI ACM Special Interest Group on Artificial Intelligence"]},"container-title":["Proceedings of the 2023 AAAI\/ACM Conference on AI, Ethics, and Society"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3600211.3604690","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3600211.3604690","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:39Z","timestamp":1750178259000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3600211.3604690"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,8]]},"references-count":65,"alternative-id":["10.1145\/3600211.3604690","10.1145\/3600211"],"URL":"https:\/\/doi.org\/10.1145\/3600211.3604690","relation":{},"subject":[],"published":{"date-parts":[[2023,8,8]]},"assertion":[{"value":"2023-08-29","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}