{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,24]],"date-time":"2025-06-24T11:40:05Z","timestamp":1750765205637,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":107,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,6,23]]},"DOI":"10.1145\/3715275.3732201","type":"proceedings-article","created":{"date-parts":[[2025,6,23]],"date-time":"2025-06-23T17:01:18Z","timestamp":1750698078000},"page":"3148-3161","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["In Which Areas of Technical AI Safety Could Geopolitical Rivals Cooperate?"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-5552-2961","authenticated-orcid":false,"given":"Ben","family":"Bucknall","sequence":"first","affiliation":[{"name":"Department of Engineering Science, University of Oxford, Oxford, United Kingdom and Oxford Martin AI Governance Initiative, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-6378-5086","authenticated-orcid":false,"given":"Saad","family":"Siddiqui","sequence":"additional","affiliation":[{"name":"Safe AI Forum, London, United Kingdom and Oxford Martin AI Governance Initiative, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-8358-3413","authenticated-orcid":false,"given":"Lara","family":"Thurnherr","sequence":"additional","affiliation":[{"name":"King's College London, London, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-3871-7371","authenticated-orcid":false,"given":"Conor","family":"McGurk","sequence":"additional","affiliation":[{"name":"Safe AI Forum, Toronto, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1813-7772","authenticated-orcid":false,"given":"Ben","family":"Harack","sequence":"additional","affiliation":[{"name":"Oxford Martin AI Governance Initiative, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7913-9296","authenticated-orcid":false,"given":"Anka","family":"Reuel","sequence":"additional","affiliation":[{"name":"Stanford University, Stanford, USA and The Belfer Center for Science and International Affairs, Harvard Kennedy School, Cambridge, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8961-9859","authenticated-orcid":false,"given":"Patricia","family":"Paskov","sequence":"additional","affiliation":[{"name":"RAND, Santa Monica, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5516-8826","authenticated-orcid":false,"given":"Casey","family":"Mahoney","sequence":"additional","affiliation":[{"name":"RAND, Santa Monica, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0315-9821","authenticated-orcid":false,"given":"S\u00f6ren","family":"Mindermann","sequence":"additional","affiliation":[{"name":"Mila - Quebec AI Institute, Montreal, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9289-378X","authenticated-orcid":false,"given":"Scott","family":"Singer","sequence":"additional","affiliation":[{"name":"Carnegie Endowment for International Peace, Washington, DC, USA and Oxford Martin AI Governance Initiative, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-2759-932X","authenticated-orcid":false,"given":"Vinay","family":"Hiremath","sequence":"additional","affiliation":[{"name":"Centre for the Governance of AI, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3952-3637","authenticated-orcid":false,"given":"Charbel-Rapha\u00ebl","family":"Segerie","sequence":"additional","affiliation":[{"name":"Centre pour la S\u00e9curit\u00e9 de l'IA (CeSIA), Paris, France"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8792-0744","authenticated-orcid":false,"given":"Oscar","family":"Delaney","sequence":"additional","affiliation":[{"name":"Institute for AI Policy and Strategy, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5627-9093","authenticated-orcid":false,"given":"Alessandro","family":"Abate","sequence":"additional","affiliation":[{"name":"Department of Computer Science, University of Oxford, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-1889-6577","authenticated-orcid":false,"given":"Fazl","family":"Barez","sequence":"additional","affiliation":[{"name":"University of Oxford, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1749-875X","authenticated-orcid":false,"given":"Michael K.","family":"Cohen","sequence":"additional","affiliation":[{"name":"University of California, Berkeley, Berkeley, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-0259-5732","authenticated-orcid":false,"given":"Philip","family":"Torr","sequence":"additional","affiliation":[{"name":"University of Oxford, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4988-1430","authenticated-orcid":false,"given":"Ferenc","family":"Husz\u00e1r","sequence":"additional","affiliation":[{"name":"University of Cambridge, Cambridge, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2082-734X","authenticated-orcid":false,"given":"Anisoara","family":"Calinescu","sequence":"additional","affiliation":[{"name":"Department of Computer Science, University of Oxford, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4070-4814","authenticated-orcid":false,"given":"Gabriel Davis","family":"Jones","sequence":"additional","affiliation":[{"name":"Oxford Digital Health Labs, University of Oxford, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6201-2811","authenticated-orcid":false,"given":"Robert","family":"Trager","sequence":"additional","affiliation":[{"name":"Oxford Martin AI Governance Initiative, Oxford, United Kingdom and Blavatnik School of Government, University of Oxford, Oxford, United Kingdom"}]}],"member":"320","published-online":{"date-parts":[[2025,6,23]]},"reference":[{"key":"e_1_3_3_2_2_2","unstructured":"2012. ISO\/IEC 19790:2012. https:\/\/www.iso.org\/standard\/52906.html"},{"key":"e_1_3_3_2_3_2","unstructured":"2023. Dealing with Risks in International Research Cooperation: Recommendations from the Deutsche Forschungsgemeinschaft. https:\/\/www.dfg.de\/resource\/blob\/289704\/585cb3b48bb8e9f5b6e57e0e0a0d700e\/risiken-int-kooperationen-en-data.pdf"},{"key":"e_1_3_3_2_4_2","unstructured":"2024. REGULATION (EU) 2024\/1689 OF THE EUROPEAN PARLIAMENT AND OF THE COUNCIL of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300\/2008 (EU) No 167\/2013 (EU) No 168\/2013 (EU) 2018\/858 (EU) 2018\/1139 and (EU) 2019\/2144 and Directives 2014\/90\/EU (EU) 2016\/797 and (EU) 2020\/1828 (Artificial Intelligence Act). https:\/\/eur-lex.europa.eu\/legal-content\/EN\/TXT\/PDF\/?uri=OJ:L_202401689"},{"volume-title":"US AISI and UK AISI Joint Pre-Deployment Test: OpenAI o1","year":"2024","key":"e_1_3_3_2_5_2","unstructured":"2024. US AISI and UK AISI Joint Pre-Deployment Test: OpenAI o1. Technical Report. https:\/\/www.nist.gov\/system\/files\/documents\/2024\/12\/18\/US_UK_AI%20Safety%20Institute_%20December_Publication-OpenAIo1.pdf"},{"key":"e_1_3_3_2_6_2","volume-title":"Secure, Governable Chips: Using On-Chip Mechanisms to Manage National Security Risks from AI & Advanced Computing","author":"Aarne Onni","year":"2024","unstructured":"Onni Aarne, Tim Fist, and Caleb Withers. 2024. Secure, Governable Chips: Using On-Chip Mechanisms to Manage National Security Risks from AI & Advanced Computing. Technical Report. Center for a New American Security. https:\/\/www.cnas.org\/publications\/reports\/secure-governable-chips"},{"key":"e_1_3_3_2_7_2","volume-title":"Voice and Access in AI: Global AI Majority Participation in Artificial Intelligence Development and Governance","author":"Adan Sumaya\u00a0N.","year":"2024","unstructured":"Sumaya\u00a0N. Adan, Robert Trager, Kayla Blomquist, Claire Dennis, Gemma Edom, Lucia Velasco, Cecil Abungu, Ben Garfinkel, Julian Jacobs, Chinasa\u00a0T. Okolo, Boxi Wu, and Jai Vipra. 2024. Voice and Access in AI: Global AI Majority Participation in Artificial Intelligence Development and Governance. Technical Report. Oxford Martin AI Governance Initiative, Oxford, UK. https:\/\/www.oxfordmartin.ox.ac.uk\/publications\/voice-and-access-in-ai-global-ai-majority-participation-in-artificial-intelligence-development-and-governance"},{"key":"e_1_3_3_2_8_2","doi-asserted-by":"publisher","unstructured":"Jide Alaga Jonas Schuett and Markus Anderljung. 2024. A Grading Rubric for AI Safety Frameworks. https:\/\/doi.org\/10.48550\/arXiv.2409.08751 arXiv:https:\/\/arXiv.org\/abs\/2409.08751 [cs].","DOI":"10.48550\/arXiv.2409.08751"},{"key":"e_1_3_3_2_9_2","unstructured":"Anthropic. 2024. Responsible Scaling Policy. https:\/\/assets.anthropic.com\/m\/24a47b00f10301cd\/original\/Anthropic-Responsible-Scaling-Policy-2024-10-15.pdf"},{"key":"e_1_3_3_2_10_2","unstructured":"Arms\u00a0Control Association. 2021. The Open Skies Treaty at a Glance. https:\/\/www.armscontrol.org\/factsheets\/openskies"},{"key":"e_1_3_3_2_11_2","volume-title":"Principles of Model Checking","author":"Baier Christel","year":"2008","unstructured":"Christel Baier and Joost-Pieter Katoen. 2008. Principles of Model Checking. The MIT Press. https:\/\/mitpress.mit.edu\/9780262026499\/principles-of-model-checking\/"},{"key":"e_1_3_3_2_12_2","volume-title":"International AI Safety Report: The International Scientific Report on the Safety of Advanced AI","author":"Bengio Yoshua","year":"2025","unstructured":"Yoshua Bengio, S\u00f6ren Mindermann, Daniel Privitera, Tamay Besiroglu, Rishi Bommasani, Stephen Casper, Yejin Choi, Philip Fox, Ben Garfinkel, Danielle Goldfarb, Hoda Heidari, Anson Ho, Sayash Kapoor, Leila Khalatbari, Shayne Longpre, Sam Manning, Vasilios Mavroudis, Mantas Mazeika, Julian Michael, Jessica Newman, Kwan\u00a0Yee Ng, Chinasa\u00a0T. Okolo, Deborah Raji, Girish Sastry, Elizabeth Seger, Theodora Skeadas, Tobin South, Emma Strubell, Florian Tram\u00e8r, Lucia Velasco, and Nicole Wheeler. 2025. International AI Safety Report: The International Scientific Report on the Safety of Advanced AI. Technical Report. Department for Science, Innovation & Technology. https:\/\/www.gov.uk\/government\/publications\/international-ai-safety-report-2025"},{"key":"e_1_3_3_2_13_2","volume-title":"International Scientific Report on the Safety of Advanced AI - Interim Report","author":"Bengio Yoshua","year":"2024","unstructured":"Yoshua Bengio, S\u00f6ren Mindermann, Daniel Privitera, Rishi Bommasani, Stephen Casper, Yejin Choi, Danielle Goldfarb, Hoda Heidari, Leila Khalatbari, Shayne Longpre, Vasilios Mavroudis, Mantas Mazeika, Kwan\u00a0Yee Ng, Chinasa\u00a0T. Okolo, Deborah Raji, Theodora Skeadas, and Florian Tram\u00e8r. 2024. International Scientific Report on the Safety of Advanced AI - Interim Report. Technical Report. Department for Science, Innovation & Technology. https:\/\/www.gov.uk\/government\/publications\/international-scientific-report-on-the-safety-of-advanced-ai"},{"key":"e_1_3_3_2_14_2","unstructured":"Emily Benson. 2023. Updated October 7 Semiconductor Export Controls. https:\/\/www.csis.org\/analysis\/updated-october-7-semiconductor-export-controls"},{"key":"e_1_3_3_2_15_2","doi-asserted-by":"publisher","unstructured":"Tamay Besiroglu Sage\u00a0Andrus Bergerson Amelia Michael Lennart Heim Xueyun Luo and Neil Thompson. 2024. The Compute Divide in Machine Learning: A Threat to Academic Contribution and Scrutiny?https:\/\/doi.org\/10.48550\/arXiv.2401.02452 arXiv:https:\/\/arXiv.org\/abs\/2401.02452 [cs].","DOI":"10.48550\/arXiv.2401.02452"},{"key":"e_1_3_3_2_16_2","doi-asserted-by":"publisher","unstructured":"Fateh Boudardara Abderraouf Boussif Pierre-Jean Meyer and Mohamed Ghazel. 2024. A Review of Abstraction Methods Toward Verifying Neural Networks. ACM Trans. Embed. Comput. Syst. 23 4 (June 2024) 58:1\u201358:19. https:\/\/doi.org\/10.1145\/3617508","DOI":"10.1145\/3617508"},{"key":"e_1_3_3_2_17_2","doi-asserted-by":"publisher","unstructured":"Miles Brundage Shahar Avin Jasmine Wang Haydn Belfield Gretchen Krueger Gillian Hadfield Heidy Khlaaf Jingying Yang Helen Toner Ruth Fong Tegan Maharaj Pang\u00a0Wei Koh Sara Hooker Jade Leung Andrew Trask Emma Bluemke Jonathan Lebensold Cullen O\u2019Keefe Mark Koren Th\u00e9o Ryffel J.\u00a0B. Rubinovitz Tamay Besiroglu Federica Carugati Jack Clark Peter Eckersley Sarah\u00a0de Haas Maritza Johnson Ben Laurie Alex Ingerman Igor Krawczuk Amanda Askell Rosario Cammarota Andrew Lohn David Krueger Charlotte Stix Peter Henderson Logan Graham Carina Prunkl Bianca Martin Elizabeth Seger Noa Zilberman Se\u00e1n\u00a0\u00d3 h\u00c9igeartaigh Frens Kroeger Girish Sastry Rebecca Kagan Adrian Weller Brian Tse Elizabeth Barnes Allan Dafoe Paul Scharre Ariel Herbert-Voss Martijn Rasser Shagun Sodhani Carrick Flynn Thomas\u00a0Krendl Gilbert Lisa Dyer Saif Khan Yoshua Bengio and Markus Anderljung. 2020. Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims. https:\/\/doi.org\/10.48550\/arXiv.2004.07213 arXiv:https:\/\/arXiv.org\/abs\/2004.07213 [cs].","DOI":"10.48550\/arXiv.2004.07213"},{"key":"e_1_3_3_2_18_2","doi-asserted-by":"publisher","unstructured":"John Burden. 2024. Evaluating AI Evaluation: Perils and Prospects. https:\/\/doi.org\/10.48550\/arXiv.2407.09221 arXiv:https:\/\/arXiv.org\/abs\/2407.09221 [cs] version: 1.","DOI":"10.48550\/arXiv.2407.09221"},{"key":"e_1_3_3_2_19_2","unstructured":"CAICT. 2024. Protecting AI security and building a model of industry self-discipline - the first batch of 17 companies signed the \"Artificial Intelligence Security Commitment\". https:\/\/mp.weixin.qq.com\/s\/s-XFKQCWhu0uye4opgb3Ng"},{"key":"e_1_3_3_2_20_2","doi-asserted-by":"publisher","unstructured":"Stephen Casper Xander Davies Claudia Shi Thomas\u00a0Krendl Gilbert J\u00e9r\u00e9my Scheurer Javier Rando Rachel Freedman Tomasz Korbak David Lindner Pedro Freire Tony Wang Samuel Marks Charbel-Rapha\u00ebl Segerie Micah Carroll Andi Peng Phillip Christoffersen Mehul Damani Stewart Slocum Usman Anwar Anand Siththaranjan Max Nadeau Eric\u00a0J. Michaud Jacob Pfau Dmitrii Krasheninnikov Xin Chen Lauro Langosco Peter Hase Erdem B\u0131y\u0131k Anca Dragan David Krueger Dorsa Sadigh and Dylan Hadfield-Menell. 2023. Open Problems and Fundamental Limitations of Reinforcement Learning from Human Feedback. https:\/\/doi.org\/10.48550\/arXiv.2307.15217 arXiv:https:\/\/arXiv.org\/abs\/2307.15217 [cs].","DOI":"10.48550\/arXiv.2307.15217"},{"key":"e_1_3_3_2_21_2","doi-asserted-by":"publisher","unstructured":"Xavier Casta\u00f1er and Nuno Oliveira. 2020. Collaboration Coordination and Cooperation Among Organizations: Establishing the Distinctive Meanings of These Terms Through a Systematic Literature Review. Journal of Management 46 6 (July 2020) 965\u20131001. https:\/\/doi.org\/10.1177\/0149206320901565 Publisher: SAGE Publications Inc.","DOI":"10.1177\/0149206320901565"},{"key":"e_1_3_3_2_22_2","volume-title":"Tech-Politik: Historical Perspectives on Innovation, Technology, and Strategic Competition","author":"Center Seth","year":"2019","unstructured":"Seth Center and Emma Bates. 2019. Tech-Politik: Historical Perspectives on Innovation, Technology, and Strategic Competition. Technical Report. Center for Strategic & International Studies. https:\/\/www.csis.org\/analysis\/tech-politik-historical-perspectives-innovation-technology-and-strategic-competition"},{"key":"e_1_3_3_2_23_2","doi-asserted-by":"publisher","unstructured":"Yupeng Chang Xu Wang Jindong Wang Yuan Wu Linyi Yang Kaijie Zhu Hao Chen Xiaoyuan Yi Cunxiang Wang Yidong Wang Wei Ye Yue Zhang Yi Chang Philip\u00a0S. Yu Qiang Yang and Xing Xie. 2024. A Survey on Evaluation of Large Language Models. ACM Trans. Intell. Syst. Technol. 15 3 (March 2024) 39:1\u201339:45. https:\/\/doi.org\/10.1145\/3641289","DOI":"10.1145\/3641289"},{"key":"e_1_3_3_2_24_2","doi-asserted-by":"publisher","unstructured":"Dami Choi Yonadav Shavit and David Duvenaud. 2023. Tools for Verifying Neural Models\u2019 Training Data. https:\/\/doi.org\/10.48550\/arXiv.2307.00682 arXiv:https:\/\/arXiv.org\/abs\/2307.00682 [cs].","DOI":"10.48550\/arXiv.2307.00682"},{"key":"e_1_3_3_2_25_2","unstructured":"Cathleen\u00a0D Cimino-Isaacs and Karen\u00a0M Sutter. 2024. Committee on Foreign Investment in the United States (CFIUS). https:\/\/crsreports.congress.gov\/product\/pdf\/IF\/IF10177"},{"key":"e_1_3_3_2_26_2","doi-asserted-by":"crossref","unstructured":"Daniel Clery. 2024. Giant fusion project is in big trouble: ITER operations delayed to 2034 with energy-producing reactions expected 5 years later. Science 385 6704 (July 2024) 10\u201311. https:\/\/www.science.org\/content\/article\/giant-international-fusion-project-big-trouble","DOI":"10.1126\/science.adr4532"},{"key":"e_1_3_3_2_27_2","doi-asserted-by":"publisher","unstructured":"Andrew\u00a0J. Coe and Jane Vaynman. 2020. Why Arms Control Is So Rare. American Political Science Review 114 2 (May 2020) 342\u2013355. https:\/\/doi.org\/10.1017\/S000305541900073X","DOI":"10.1017\/S000305541900073X"},{"key":"e_1_3_3_2_28_2","doi-asserted-by":"publisher","unstructured":"Josef Dai Xuehai Pan Ruiyang Sun Jiaming Ji Xinbo Xu Mickel Liu Yizhou Wang and Yaodong Yang. 2023. Safe RLHF: Safe Reinforcement Learning from Human Feedback. https:\/\/doi.org\/10.48550\/arXiv.2310.12773 arXiv:https:\/\/arXiv.org\/abs\/2310.12773 [cs].","DOI":"10.48550\/arXiv.2310.12773"},{"key":"e_1_3_3_2_29_2","volume-title":"Safeguarded AI: Constructing guaranteed safety","author":"Dalrymple David\u00a0\"davidad\"","year":"2024","unstructured":"David\u00a0\"davidad\" Dalrymple. 2024. Safeguarded AI: Constructing guaranteed safety. Technical Report. Advanced Research + Invention Agency. https:\/\/www.aria.org.uk\/media\/3nhijno4\/aria-safeguarded-ai-programme-thesis-v1.pdf"},{"key":"e_1_3_3_2_30_2","doi-asserted-by":"publisher","unstructured":"David\u00a0\"davidad\" Dalrymple Joar Skalse Yoshua Bengio Stuart Russell Max Tegmark Sanjit Seshia Steve Omohundro Christian Szegedy Ben Goldhaber Nora Ammann Alessandro Abate Joe Halpern Clark Barrett Ding Zhao Tan Zhi-Xuan Jeannette Wing and Joshua Tenenbaum. 2024. Towards Guaranteed Safe AI: A Framework for Ensuring Robust and Reliable AI Systems. https:\/\/doi.org\/10.48550\/arXiv.2405.06624 arXiv:https:\/\/arXiv.org\/abs\/2405.06624 [cs].","DOI":"10.48550\/arXiv.2405.06624"},{"key":"e_1_3_3_2_31_2","doi-asserted-by":"publisher","unstructured":"Sumanth Dathathri Abigail See Sumedh Ghaisas Po-Sen Huang Rob McAdam Johannes Welbl Vandana Bachani Alex Kaskasoli Robert Stanforth Tatiana Matejovicova Jamie Hayes Nidhi Vyas Majd\u00a0Al Merey Jonah Brown-Cohen Rudy Bunel Borja Balle Taylan Cemgil Zahra Ahmed Kitty Stacpoole Ilia Shumailov Ciprian Baetu Sven Gowal Demis Hassabis and Pushmeet Kohli. 2024. Scalable watermarking for identifying large language model outputs. Nature 634 8035 (Oct. 2024) 818\u2013823. https:\/\/doi.org\/10.1038\/s41586-024-08025-4 Publisher: Nature Publishing Group.","DOI":"10.1038\/s41586-024-08025-4"},{"key":"e_1_3_3_2_32_2","unstructured":"Google DeepMind. 2024. Frontier Safety Framework v1.0."},{"key":"e_1_3_3_2_33_2","doi-asserted-by":"publisher","unstructured":"Jeffrey Ding. 2024. Keep your enemies safer: technical cooperation and transferring nuclear safety and security technologies. European Journal of International Relations 30 4 (Dec. 2024) 918\u2013945. https:\/\/doi.org\/10.1177\/13540661241246622 Publisher: SAGE Publications Ltd.","DOI":"10.1177\/13540661241246622"},{"key":"e_1_3_3_2_34_2","doi-asserted-by":"publisher","unstructured":"Diego Dorn Alexandre Variengien Charbel-Rapha\u00ebl Segerie and Vincent Corruble. 2024. BELLS: A Framework Towards Future Proof Benchmarks for the Evaluation of LLM Safeguards. https:\/\/doi.org\/10.48550\/arXiv.2406.01364 arXiv:https:\/\/arXiv.org\/abs\/2406.01364 [cs].","DOI":"10.48550\/arXiv.2406.01364"},{"key":"e_1_3_3_2_35_2","doi-asserted-by":"publisher","unstructured":"Michael Feffer Anusha Sinha Wesley\u00a0Hanwen Deng Zachary\u00a0C. Lipton and Hoda Heidari. 2024. Red-Teaming for Generative AI: Silver Bullet or Security Theater?https:\/\/doi.org\/10.48550\/arXiv.2401.15897 arXiv:https:\/\/arXiv.org\/abs\/2401.15897 [cs].","DOI":"10.48550\/arXiv.2401.15897"},{"key":"e_1_3_3_2_36_2","doi-asserted-by":"publisher","unstructured":"Jaden Fiotto-Kaufman Alexander\u00a0R. Loftus Eric Todd Jannik Brinkmann Koyena Pal Dmitrii Troitskii Michael Ripa Adam Belfki Can Rager Caden Juang Aaron Mueller Samuel Marks Arnab\u00a0Sen Sharma Francesca Lucchetti Nikhil Prakash Carla Brodley Arjun Guha Jonathan Bell Byron\u00a0C. Wallace and David Bau. 2025. NNsight and NDIF: Democratizing Access to Open-Weight Foundation Model Internals. https:\/\/doi.org\/10.48550\/arXiv.2407.14561 arXiv:https:\/\/arXiv.org\/abs\/2407.14561 [cs].","DOI":"10.48550\/arXiv.2407.14561"},{"volume-title":"FLI AI Safety Index 2024: Independent experts evaluate safety practices of leading AI companies across critical domains","year":"2024","key":"e_1_3_3_2_37_2","unstructured":"FLI. 2024. FLI AI Safety Index 2024: Independent experts evaluate safety practices of leading AI companies across critical domains. Technical Report. Future of Life Institute. http:\/\/futureoflife.org\/index"},{"key":"e_1_3_3_2_38_2","unstructured":"Center for Arms Control and Non-Proliferation. 2017. Fact Sheet: The Threshold Test Ban Treaty (TTBT). https:\/\/armscontrolcenter.org\/fact-sheet-threshold-test-ban-treaty-ttbt\/"},{"key":"e_1_3_3_2_39_2","unstructured":"Department for Science Innovation and\u00a0Technology. 2023. The Bletchley Declaration by Countries Attending the AI Safety Summit 1-2 November 2023. https:\/\/www.gov.uk\/government\/publications\/ai-safety-summit-2023-the-bletchley-declaration\/the-bletchley-declaration-by-countries-attending-the-ai-safety-summit-1-2-november-2023"},{"key":"e_1_3_3_2_40_2","unstructured":"Department for Science Innovation and\u00a0Technology. 2024. Frontier AI Safety Commitments AI Seoul Summit 2024. https:\/\/www.gov.uk\/government\/publications\/frontier-ai-safety-commitments-ai-seoul-summit-2024\/frontier-ai-safety-commitments-ai-seoul-summit-2024"},{"key":"e_1_3_3_2_41_2","unstructured":"Department for Science Innovation and\u00a0Technology. 2024. Seoul Ministerial Statement for advancing AI safety innovation and inclusivity: AI Seoul Summit 2024. https:\/\/www.gov.uk\/government\/publications\/seoul-ministerial-statement-for-advancing-ai-safety-innovation-and-inclusivity-ai-seoul-summit-2024\/seoul-ministerial-statement-for-advancing-ai-safety-innovation-and-inclusivity-ai-seoul-summit-2024"},{"key":"e_1_3_3_2_42_2","doi-asserted-by":"publisher","unstructured":"Nancy\u00a0W. Gallagher. 1997. The politics of verification: Why \u2018how much?\u2019 Is not enough. Contemporary Security Policy 18 2 (Aug. 1997) 138\u2013170. https:\/\/doi.org\/10.1080\/13523269708404165 Publisher: Routledge.","DOI":"10.1080\/13523269708404165"},{"key":"e_1_3_3_2_43_2","doi-asserted-by":"crossref","unstructured":"Sanjam Garg Aarushi Goel Somesh Jha Saeed Mahloujifar Mohammad Mahmoody Guru-Vamsi Policharla and Mingyuan Wang. 2023. Experimenting with Zero-Knowledge Proofs of Training. https:\/\/eprint.iacr.org\/2023\/1345 Publication info: Published elsewhere. Major revision. ACM CCS 2023.","DOI":"10.1145\/3576915.3623202"},{"key":"e_1_3_3_2_44_2","unstructured":"Soumya\u00a0Suvra Ghosal Souradip Chakraborty Jonas Geiping Furong Huang Dinesh Manocha and Amrit Bedi. 2023. A Survey on the Possibilities & Impossibilities of AI-generated Text Detection. Transactions on Machine Learning Research (Oct. 2023). https:\/\/openreview.net\/forum?id=AXtFeYjboj"},{"key":"e_1_3_3_2_45_2","doi-asserted-by":"publisher","unstructured":"Ryan Greenblatt Buck Shlegeris Kshitij Sachan and Fabien Roger. 2024. AI Control: Improving Safety Despite Intentional Subversion. https:\/\/doi.org\/10.48550\/arXiv.2312.06942 arXiv:https:\/\/arXiv.org\/abs\/2312.06942 [cs].","DOI":"10.48550\/arXiv.2312.06942"},{"key":"e_1_3_3_2_46_2","doi-asserted-by":"publisher","unstructured":"Charlie Griffin Louis Thomson Buck Shlegeris and Alessandro Abate. 2024. Games for AI Control: Models of Safety Evaluations of AI Deployment Protocols. https:\/\/doi.org\/10.48550\/arXiv.2409.07985 arXiv:https:\/\/arXiv.org\/abs\/2409.07985 [cs].","DOI":"10.48550\/arXiv.2409.07985"},{"key":"e_1_3_3_2_47_2","volume-title":"Safeguarding the Safeguards: How best to promote AI alignment in the public interest","author":"Guest Oliver","year":"2023","unstructured":"Oliver Guest, Michael Aird, and Se\u00e1n \u00d3\u00a0h\u00c9igeartaigh. 2023. Safeguarding the Safeguards: How best to promote AI alignment in the public interest. Technical Report. Institute for AI Policy and Strategy. https:\/\/www.iaps.ai\/research\/safeguarding-the-safeguards"},{"key":"e_1_3_3_2_48_2","volume-title":"Topics for track IIs: What can be discussed in dialogues about advanced AI risks without leaking sensitive information?","author":"Guest Oliver","year":"2024","unstructured":"Oliver Guest and Zoe Williams. 2024. Topics for track IIs: What can be discussed in dialogues about advanced AI risks without leaking sensitive information? Technical Report. Institute for AI Policy and Strategy. https:\/\/static1.squarespace.com\/static\/64edf8e7f2b10d716b5ba0e1\/t\/6633b93601a0553b73d56095\/1714665783885\/%5BFinal%5D+Topics+for+track+IIs.pdf"},{"key":"e_1_3_3_2_49_2","doi-asserted-by":"publisher","unstructured":"Kaiming He Xiangyu Zhang Shaoqing Ren and Jian Sun. 2015. Deep Residual Learning for Image Recognition. https:\/\/doi.org\/10.48550\/arXiv.1512.03385 arXiv:https:\/\/arXiv.org\/abs\/1512.03385 [cs].","DOI":"10.48550\/arXiv.1512.03385"},{"key":"e_1_3_3_2_50_2","doi-asserted-by":"publisher","DOI":"10.1201\/9781003530336"},{"key":"e_1_3_3_2_51_2","doi-asserted-by":"publisher","unstructured":"Dan Hendrycks Mantas Mazeika and Thomas Woodside. 2023. An Overview of Catastrophic AI Risks. https:\/\/doi.org\/10.48550\/arXiv.2306.12001 arXiv:https:\/\/arXiv.org\/abs\/2306.12001 [cs].","DOI":"10.48550\/arXiv.2306.12001"},{"key":"e_1_3_3_2_52_2","volume-title":"U.S. Export Controls and Economic Sanctions (fourth edition, fourth edition ed.)","author":"Hirschhorn Eric\u00a0L.","year":"2022","unstructured":"Eric\u00a0L. Hirschhorn, Brian\u00a0J. Egan, Edward\u00a0J. Krauland, Eric\u00a0L. Hirschhorn, Brian\u00a0J. Egan, and Edward\u00a0J. Krauland. 2022. U.S. Export Controls and Economic Sanctions (fourth edition, fourth edition ed.). Oxford University Press, Oxford, New York."},{"key":"e_1_3_3_2_53_2","doi-asserted-by":"publisher","unstructured":"C.\u00a0A.\u00a0R. Hoare. 1969. An axiomatic basis for computer programming. Commun. ACM 12 10 (Oct. 1969) 576\u2013580. https:\/\/doi.org\/10.1145\/363235.363259","DOI":"10.1145\/363235.363259"},{"key":"e_1_3_3_2_54_2","unstructured":"The\u00a0White House. 2022. Blueprint for an AI Bill of Rights: Making automated systems work for the American people. https:\/\/www.whitehouse.gov\/ostp\/ai-bill-of-rights\/"},{"key":"e_1_3_3_2_55_2","unstructured":"Yukon Huang Isaac\u00a0B. Kardon and Matt Sheehan. 2023. Three Takeaways From the Biden-Xi Meeting. https:\/\/carnegieendowment.org\/posts\/2023\/11\/three-takeaways-from-the-biden-xi-meeting?lang=en"},{"key":"e_1_3_3_2_56_2","unstructured":"IAEA. 2014. IAEA Safeguards Overview. https:\/\/www.iaea.org\/publications\/factsheets\/iaea-safeguards-overview Publisher: IAEA."},{"key":"e_1_3_3_2_57_2","unstructured":"UK\u00a0AI\u00a0Safety Institute. 2024. Conference on frontier AI safety frameworks. https:\/\/www.aisi.gov.uk\/work\/conference-on-frontier-ai-safety-frameworks"},{"key":"e_1_3_3_2_58_2","unstructured":"UK\u00a0AI\u00a0Safety Institute. 2024. Early lessons from evaluating frontier AI systems. https:\/\/www.aisi.gov.uk\/work\/early-lessons-from-evaluating-frontier-ai-systems"},{"key":"e_1_3_3_2_59_2","doi-asserted-by":"publisher","unstructured":"Robert Jervis. 1978. Cooperation Under the Security Dilemma. World Politics 30 2 (1978) 167\u2013214. https:\/\/doi.org\/10.2307\/2009958 Publisher: [Trustees of Princeton University The Johns Hopkins University Press].","DOI":"10.2307\/2009958"},{"key":"e_1_3_3_2_60_2","unstructured":"Wang Jingjing. 2016. The Whampoa Academy of China\u2019s Internet. https:\/\/weibo.com\/p\/1001643998598932131471 English commentary and translation by Jeffrey Ding available at https:\/\/chinai.substack.com\/p\/chinai-37-happy-20th-anniversary."},{"key":"e_1_3_3_2_61_2","doi-asserted-by":"publisher","unstructured":"Daniel Kang Tatsunori Hashimoto Ion Stoica and Yi Sun. 2022. Scaling up Trustless DNN Inference with Zero-Knowledge Proofs. https:\/\/doi.org\/10.48550\/arXiv.2210.08674 arXiv:https:\/\/arXiv.org\/abs\/2210.08674 [cs].","DOI":"10.48550\/arXiv.2210.08674"},{"key":"e_1_3_3_2_62_2","unstructured":"Holden Karnofsky. 2024. If-Then Commitments for AI Risk Reduction. https:\/\/carnegieendowment.org\/research\/2024\/09\/if-then-commitments-for-ai-risk-reduction?lang=en"},{"key":"e_1_3_3_2_63_2","doi-asserted-by":"publisher","unstructured":"Leonie Koessler Jonas Schuett and Markus Anderljung. 2024. Risk thresholds for frontier AI. https:\/\/doi.org\/10.48550\/arXiv.2406.14713 arXiv:https:\/\/arXiv.org\/abs\/2406.14713 [cs].","DOI":"10.48550\/arXiv.2406.14713"},{"key":"e_1_3_3_2_64_2","doi-asserted-by":"publisher","unstructured":"Stephen\u00a0D. Krasner. 1991. Global Communications and National Power: Life on the Pareto Frontier. World Politics 43 3 (April 1991) 336\u2013366. https:\/\/doi.org\/10.2307\/2010398","DOI":"10.2307\/2010398"},{"key":"e_1_3_3_2_65_2","volume-title":"Hardware-Enabled Governance Mechanisms: Developing Technical Solutions to Exempt Items Otherwise Classified Under Export Control Classification Numbers 3A090 and 4A090","author":"Kulp Gabriel","year":"2024","unstructured":"Gabriel Kulp, Daniel Gonzales, Everett Smith, Lennart Heim, Prateek Puri, Michael J.\u00a0D. Vermeer, and Zev Winkelman. 2024. Hardware-Enabled Governance Mechanisms: Developing Technical Solutions to Exempt Items Otherwise Classified Under Export Control Classification Numbers 3A090 and 4A090. Technical Report. RAND Corporation. https:\/\/www.rand.org\/pubs\/working_papers\/WRA3056-1.html"},{"key":"e_1_3_3_2_66_2","doi-asserted-by":"publisher","unstructured":"Nancy\u00a0G. Leveson and John\u00a0P. Thomas. 2023. Certification of Safety-Critical Systems. Commun. ACM 66 10 (Sept. 2023) 22\u201326. https:\/\/doi.org\/10.1145\/3615860","DOI":"10.1145\/3615860"},{"key":"e_1_3_3_2_67_2","unstructured":"Michael Martina and Trevor Hunnicutt. 2024. US China meet in Geneva to discuss AI risks. Reuters (May 2024). https:\/\/www.reuters.com\/technology\/us-china-meet-geneva-discuss-ai-risks-2024-05-13\/"},{"key":"e_1_3_3_2_68_2","doi-asserted-by":"publisher","DOI":"10.1038\/462412a"},{"key":"e_1_3_3_2_69_2","volume-title":"Chip War: The Fight for the World\u2019s Most Critical Technology","author":"Miller Chris","year":"2023","unstructured":"Chris Miller. 2023. Chip War: The Fight for the World\u2019s Most Critical Technology. Simon & Schuster. https:\/\/www.simonandschuster.co.uk\/books\/Chip-War\/Chris-Miller\/9781398504127"},{"key":"e_1_3_3_2_70_2","volume-title":"The Operational Risks of AI in Large-Scale Biological Attacks: A Red-Team Approach","author":"Mouton Christopher\u00a0A.","year":"2023","unstructured":"Christopher\u00a0A. Mouton, Caleb Lucas, and Ella Guest. 2023. The Operational Risks of AI in Large-Scale Biological Attacks: A Red-Team Approach. Technical Report. RAND Corporation. https:\/\/www.rand.org\/pubs\/research_reports\/RRA2977-1.html"},{"key":"e_1_3_3_2_71_2","doi-asserted-by":"publisher","DOI":"10.6028\/NIST.AI.600-1"},{"key":"e_1_3_3_2_72_2","doi-asserted-by":"publisher","unstructured":"Peter Naur. 1966. Proof of algorithms by general snapshots. BIT Numerical Mathematics 6 4 (July 1966) 310\u2013316. https:\/\/doi.org\/10.1007\/BF01966091","DOI":"10.1007\/BF01966091"},{"key":"e_1_3_3_2_73_2","volume-title":"Securing AI Model Weights: Preventing Theft and Misuse of Frontier Models","author":"Nevo Sella","year":"2024","unstructured":"Sella Nevo, Dan Lahav, Ajay Karpur, Yogev Bar-On, Henry\u00a0Alexander Bradley, and Jeff Alstott. 2024. Securing AI Model Weights: Preventing Theft and Misuse of Frontier Models. Technical Report. RAND Corporation. https:\/\/www.rand.org\/pubs\/research_reports\/RRA2849-1.html"},{"key":"e_1_3_3_2_74_2","unstructured":"Emerging\u00a0Technology Observatory. 2024. Country Activity Tracker (CAT): Artificial Intelligence. https:\/\/cat.eto.tech\/"},{"key":"e_1_3_3_2_75_2","volume-title":"Improving International Testing of Foundation Models: A Pilot Testing Exercise from the International Network of AI Safety Institutes","author":"AI\u00a0Safety\u00a0Institutes International\u00a0Network of","year":"2024","unstructured":"International\u00a0Network of AI\u00a0Safety\u00a0Institutes. 2024. Improving International Testing of Foundation Models: A Pilot Testing Exercise from the International Network of AI Safety Institutes. Technical Report. San Francisco. https:\/\/www.nist.gov\/system\/files\/documents\/2024\/11\/21\/Improving%20International%20Testing%20of%20Foundation%20Models-%20%20%20A%20Pilot%20Testing%20Exercise%20from%20the%20International%20Network%20of%20AI%20Safety%20Institutes.pdf"},{"key":"e_1_3_3_2_76_2","unstructured":"U.S.\u00a0Department of Commerce. 2024. U.S. Secretary of Commerce Raimondo and U.S. Secretary of State Blinken Announce Inaugural Convening of International Network of AI Safety Institutes in San Francisco | U.S. Department of Commerce. https:\/\/www.commerce.gov\/news\/press-releases\/2024\/09\/us-secretary-commerce-raimondo-and-us-secretary-state-blinken-announce"},{"key":"e_1_3_3_2_77_2","unstructured":"OpenAI. 2023. Preparedness Framework (Beta). https:\/\/cdn.openai.com\/openai-preparedness-framework-beta.pdf"},{"key":"e_1_3_3_2_78_2","unstructured":"World\u00a0Health Organization. 2021. Ethics and Governance of Artificial Intelligence for Health: WHO guidance. https:\/\/www.who.int\/publications\/i\/item\/9789240029200"},{"key":"e_1_3_3_2_79_2","volume-title":"Coordinated Disclosure of Dual-Use Capabilities: An Early Warning System for Advanced AI","author":"O\u2019Brien Joe","year":"2024","unstructured":"Joe O\u2019Brien, Shaun Ee, Jam Kraprayoon, Bill Anderson-Samways, Oscar Delaney, and Zoe Williams. 2024. Coordinated Disclosure of Dual-Use Capabilities: An Early Warning System for Advanced AI. Technical Report. Institute for AI Policy and Strategy. https:\/\/www.iaps.ai\/research\/coordinated-disclosure"},{"key":"e_1_3_3_2_80_2","volume-title":"How Chinese Companies Facilitate Technology Transfer from the United States","author":"O\u2019Connor Sean","year":"2019","unstructured":"Sean O\u2019Connor. 2019. How Chinese Companies Facilitate Technology Transfer from the United States. Staff Research Report. U.S.-China Economic and Security Review Commission, Washington, DC. https:\/\/www.uscc.gov\/sites\/default\/files\/Research\/How%20Chinese%20Companies%20Facilitate%20Tech%20Transfer%20from%20the%20US.pdf"},{"key":"e_1_3_3_2_81_2","doi-asserted-by":"publisher","unstructured":"Qianqian Pan Mianxiong Dong Kaoru Ota and Jun Wu. 2022. Device-Bind Key-Storageless Hardware AI Model IP Protection: A PUF and Permute-Diffusion Encryption-Enabled Approach. https:\/\/doi.org\/10.48550\/arXiv.2212.11133 arXiv:https:\/\/arXiv.org\/abs\/2212.11133 [cs].","DOI":"10.48550\/arXiv.2212.11133"},{"key":"e_1_3_3_2_82_2","volume-title":"Interim Report: Mechanisms for Flexible Hardware-Enabled Guarantees","author":"Petrie James","year":"2024","unstructured":"James Petrie, Onni Aarne, Nora Ammann, and David\u00a0\"davidad\" Dalrymple. 2024. Interim Report: Mechanisms for Flexible Hardware-Enabled Guarantees. Technical Report."},{"key":"e_1_3_3_2_83_2","volume-title":"The Future of International Scientific Assessments of AI\u2019s Risks","author":"Pouget Hadrien","year":"2024","unstructured":"Hadrien Pouget, Claire Dennis, Jon Bateman, Robert\u00a0F. Trager, Renan Araujo, Haydn Belfield, Belinda Cleeland, Malou Estier, Gideon Futerman, Oliver Guest, Carlos\u00a0Ignacio Gutierrez, Vishnu Kannan, Casey Mahoney, Matthijs Maas, Charles Martinet, Jakob M\u00f6kander, Kwan\u00a0Yee Ng, Se\u00e1n \u00d3\u00a0h\u00c9igeartaigh, Aidan Peppin, Konrad Seifert, Scott Singer, Maxime Stauffer, Caleb Withers, and Marta Ziosi. 2024. The Future of International Scientific Assessments of AI\u2019s Risks. Technical Report. Oxford Martin AI Governance Initiative, Oxford, UK. https:\/\/www.oxfordmartin.ox.ac.uk\/publications\/the-future-of-international-scientific-assessments-of-ais-risks"},{"key":"e_1_3_3_2_84_2","unstructured":"Jarrett Renshaw and Trevor Hunnicutt. 2024. Biden Xi agree that humans not AI should control nuclear arms. Reuters (Nov. 2024). https:\/\/www.reuters.com\/world\/biden-xi-agreed-that-humans-not-ai-should-control-nuclear-weapons-white-house-2024-11-16\/"},{"key":"e_1_3_3_2_85_2","unstructured":"Anka Reuel Ben Bucknall Stephen Casper Tim Fist Lisa Soder Onni Aarne Lewis Hammond Lujain Ibrahim Alan Chan Peter Wills Markus Anderljung Ben Garfinkel Lennart Heim Andrew Trask Gabriel Mukobi Rylan Schaeffer Mauricio Baker Sara Hooker Irene Solaiman Alexandra\u00a0Sasha Luccioni Nitarshan Rajkumar Nicolas Mo\u00ebs Jeffrey Ladish Neel Guha Jessica Newman Yoshua Bengio Tobin South Alex Pentland Sanmi Koyejo Mykel\u00a0J. Kochenderfer and Robert Trager. 2024. Open Problems in Technical AI Governance. http:\/\/arxiv.org\/abs\/2407.14981 arXiv:https:\/\/arXiv.org\/abs\/2407.14981 [cs]."},{"key":"e_1_3_3_2_86_2","doi-asserted-by":"publisher","unstructured":"Angelika Romanou Negar Foroutan Anna Sotnikova Zeming Chen Sree\u00a0Harsha Nelaturu Shivalika Singh Rishabh Maheshwary Micol Altomare Mohamed\u00a0A. Haggag Snegha A Alfonso Amayuelas Azril\u00a0Hafizi Amirudin Viraat Aryabumi Danylo Boiko Michael Chang Jenny Chim Gal Cohen Aditya\u00a0Kumar Dalmia Abraham Diress Sharad Duwal Daniil Dzenhaliou Daniel Fernando\u00a0Erazo Florez Fabian Farestam Joseph\u00a0Marvin Imperial Shayekh\u00a0Bin Islam Perttu Isotalo Maral Jabbarishiviari B\u00f6rje\u00a0F. Karlsson Eldar Khalilov Christopher Klamm Fajri Koto Dominik Krzemi\u0144ski Gabriel Adriano\u00a0de Melo Syrielle Montariol Yiyang Nan Joel Niklaus Jekaterina Novikova Johan Samir\u00a0Obando Ceron Debjit Paul Esther Ploeger Jebish Purbey Swati Rajwal Selvan\u00a0Sunitha Ravi Sara Rydell Roshan Santhosh Drishti Sharma Marjana\u00a0Prifti Skenduli Arshia\u00a0Soltani Moakhar Bardia\u00a0Soltani Moakhar Ran Tamir Ayush\u00a0Kumar Tarun Azmine\u00a0Toushik Wasi Thenuka\u00a0Ovin Weerasinghe Serhan Yilmaz Mike Zhang Imanol Schlag Marzieh Fadaee Sara Hooker and Antoine Bosselut. 2024. INCLUDE: Evaluating Multilingual Language Understanding with Regional Knowledge. https:\/\/doi.org\/10.48550\/arXiv.2411.19799 arXiv:https:\/\/arXiv.org\/abs\/2411.19799 [cs].","DOI":"10.48550\/arXiv.2411.19799"},{"key":"e_1_3_3_2_87_2","unstructured":"Tim R\u00fchlig. 2023. The Geopolitics of Technical Standardization. https:\/\/dgap.org\/en\/research\/publications\/geopolitics-technical-standardization"},{"key":"e_1_3_3_2_88_2","doi-asserted-by":"publisher","unstructured":"Girish Sastry Lennart Heim Haydn Belfield Markus Anderljung Miles Brundage Julian Hazell Cullen O\u2019Keefe Gillian\u00a0K. Hadfield Richard Ngo Konstantin Pilz George Gor Emma Bluemke Sarah Shoker Janet Egan Robert\u00a0F. Trager Shahar Avin Adrian Weller Yoshua Bengio and Diane Coyle. 2024. Computing Power and the Governance of Artificial Intelligence. https:\/\/doi.org\/10.48550\/arXiv.2402.08797 arXiv:https:\/\/arXiv.org\/abs\/2402.08797 [cs].","DOI":"10.48550\/arXiv.2402.08797"},{"key":"e_1_3_3_2_89_2","doi-asserted-by":"publisher","unstructured":"Jonas Schuett Markus Anderljung Alexis Carlier Leonie Koessler and Ben Garfinkel. 2024. From Principles to Rules: A Regulatory Approach for Frontier AI. https:\/\/doi.org\/10.48550\/arXiv.2407.07300 arXiv:https:\/\/arXiv.org\/abs\/2407.07300 [cs].","DOI":"10.48550\/arXiv.2407.07300"},{"key":"e_1_3_3_2_90_2","doi-asserted-by":"publisher","unstructured":"Yonadav Shavit. 2023. What does it take to catch a Chinchilla? Verifying Rules on Large-Scale Neural Network Training via Compute Monitoring. https:\/\/doi.org\/10.48550\/arXiv.2303.11341 arXiv:https:\/\/arXiv.org\/abs\/2303.11341 [cs].","DOI":"10.48550\/arXiv.2303.11341"},{"key":"e_1_3_3_2_91_2","unstructured":"Matt Sheehan and Jacob Feldgoise. 2023. What Washington Gets Wrong About China and Technical Standards. https:\/\/carnegieendowment.org\/research\/2023\/02\/what-washington-gets-wrong-about-china-and-technical-standards?lang=en"},{"key":"e_1_3_3_2_92_2","unstructured":"Scott Singer. 2024. How the UK Should Engage China at AI\u2019s Frontier. https:\/\/carnegieendowment.org\/posts\/2024\/10\/lammy-china-ai-safety-cooperation?lang=en"},{"key":"e_1_3_3_2_93_2","doi-asserted-by":"publisher","unstructured":"Tobin South Alexander Camuto Shrey Jain Shayla Nguyen Robert Mahari Christian Paquin Jason Morton and Alex\u00a0\u2019Sandy\u2019 Pentland. 2024. Verifiable evaluations of machine learning models using zkSNARKs. https:\/\/doi.org\/10.48550\/arXiv.2402.02675 arXiv:https:\/\/arXiv.org\/abs\/2402.02675 [cs].","DOI":"10.48550\/arXiv.2402.02675"},{"key":"e_1_3_3_2_94_2","doi-asserted-by":"publisher","unstructured":"Arthur\u00a0A. Stein. 1982. Coordination and collaboration: regimes in an anarchic world. International Organization 36 2 (1982) 299\u2013324. https:\/\/doi.org\/10.1017\/S0020818300018968","DOI":"10.1017\/S0020818300018968"},{"key":"e_1_3_3_2_95_2","doi-asserted-by":"publisher","unstructured":"Merlin Stein Jamie Bernardi and Connor Dunlop. 2024. The Role of Governments in Increasing Interconnected Post-Deployment Monitoring of AI. https:\/\/doi.org\/10.48550\/arXiv.2410.04931 arXiv:https:\/\/arXiv.org\/abs\/2410.04931 [cs].","DOI":"10.48550\/arXiv.2410.04931"},{"key":"e_1_3_3_2_96_2","unstructured":"Merlin Stein and Connor Dunlop. 2024. Safe beyond sale: post-deployment monitoring of AI. https:\/\/www.adalovelaceinstitute.org\/blog\/post-deployment-monitoring-of-ai\/"},{"key":"e_1_3_3_2_97_2","unstructured":"Mengqi Sun. 2024. U.S. China to Cooperate in the Fight Against Dirty Money. Wall Street Journal (April 2024). https:\/\/www.wsj.com\/articles\/u-s-china-to-cooperate-in-the-fight-against-dirty-money-1edb9a25"},{"key":"e_1_3_3_2_98_2","unstructured":"Andrew Trask Aziz\u00a0Berkay Yesilyurt Bennett Farkas Callis Ezenwaka Carmen Popa Dave Buckley Eelco van\u00a0der Wel Francesco Mosconi Grace Han Ionesio Junior Irina Bejan Ishan Mishra Khoa Nguyen Koen van\u00a0der Veen Kyoko Eng Lacey Strahm Logan Graham Madhava Jay Matei Simtinica Osam Kyemenu-Sarsah Peter Smith Rasswanth S Ronnie Falcon Sameer Wagh Sandeep Mandala Shubham Gupta Stephen Gabriel Subha Ramkumar Tauquir Ahmed Teo Milea Valerio Maggio Yash Gorana and Zarreen Reza. 2024. Secure Enclaves for AI Evaluation. https:\/\/blog.openmined.org\/secure-enclaves-for-ai-evaluation\/"},{"key":"e_1_3_3_2_99_2","unstructured":"UKRI. 2022. Managing risks in international research and innovation: An overview of higher education sector guidance. https:\/\/www.ukri.org\/wp-content\/uploads\/2022\/07\/UKRI-07072022-managing-risks-in-international-research-and-innovation-uuk-cpni-ukri_1.pdf"},{"key":"e_1_3_3_2_100_2","volume-title":"Govering AI for Humanity: Final Report","author":"HLAB UN","year":"2024","unstructured":"UN HLAB. 2024. Govering AI for Humanity: Final Report. Technical Report. United Nations, New York, NY. https:\/\/www.un.org\/ai-advisory-body"},{"key":"e_1_3_3_2_101_2","doi-asserted-by":"publisher","unstructured":"Caterina Urban and Antoine Min\u00e9. 2021. A Review of Formal Methods applied to Machine Learning. https:\/\/doi.org\/10.48550\/arXiv.2104.02466 arXiv:https:\/\/arXiv.org\/abs\/2104.02466 [cs].","DOI":"10.48550\/arXiv.2104.02466"},{"key":"e_1_3_3_2_102_2","doi-asserted-by":"publisher","unstructured":"Jason Wei Xuezhi Wang Dale Schuurmans Maarten Bosma Brian Ichter Fei Xia Ed Chi Quoc Le and Denny Zhou. 2023. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. https:\/\/doi.org\/10.48550\/arXiv.2201.11903 arXiv:https:\/\/arXiv.org\/abs\/2201.11903 [cs].","DOI":"10.48550\/arXiv.2201.11903"},{"key":"e_1_3_3_2_103_2","doi-asserted-by":"publisher","unstructured":"Christian Schroeder\u00a0de Witt Samuel Sokota J.\u00a0Zico Kolter Jakob Foerster and Martin Strohmeier. 2023. Perfectly Secure Steganography Using Minimum Entropy Coupling. https:\/\/doi.org\/10.48550\/arXiv.2210.14889 arXiv:https:\/\/arXiv.org\/abs\/2210.14889 [cs].","DOI":"10.48550\/arXiv.2210.14889"},{"key":"e_1_3_3_2_104_2","doi-asserted-by":"publisher","DOI":"10.1353\/book.66187"},{"key":"e_1_3_3_2_105_2","doi-asserted-by":"publisher","unstructured":"Andy\u00a0K. Zhang Kevin Klyman Yifan Mai Yoav Levine Yian Zhang Rishi Bommasani and Percy Liang. 2024. Language model developers should report train-test overlap. https:\/\/doi.org\/10.48550\/arXiv.2410.08385 arXiv:https:\/\/arXiv.org\/abs\/2410.08385 [cs].","DOI":"10.48550\/arXiv.2410.08385"},{"key":"e_1_3_3_2_106_2","volume-title":"The AI Index 2022 Annual Report","author":"Zhang Daniel","year":"2022","unstructured":"Daniel Zhang, Nestor Maslej, Erik Brynjolfsson, John Etchemendy, Terah Lyons, James Manyika, Helen Ngo, Juan\u00a0Carlos Niebles, Michael Sellitto, Ellie Sakhaee, Yoav Shoham, Jack Clark, and Raymond Perrault. 2022. The AI Index 2022 Annual Report. Technical Report. Stanford Institute for Human-Centered AI, Stanford University. https:\/\/aiindex.stanford.edu\/wp-content\/uploads\/2022\/03\/2022-AI-Index-Report_Master.pdf"},{"key":"e_1_3_3_2_107_2","doi-asserted-by":"publisher","unstructured":"Kun Zhou Yutao Zhu Zhipeng Chen Wentong Chen Wayne\u00a0Xin Zhao Xu Chen Yankai Lin Ji-Rong Wen and Jiawei Han. 2023. Don\u2019t Make Your LLM an Evaluation Benchmark Cheater. https:\/\/doi.org\/10.48550\/arXiv.2311.01964 arXiv:https:\/\/arXiv.org\/abs\/2311.01964 [cs].","DOI":"10.48550\/arXiv.2311.01964"},{"key":"e_1_3_3_2_108_2","doi-asserted-by":"publisher","unstructured":"Nicholas Z\u00fa\u00f1iga Saheli\u00a0Datta Burton Filippo Blancato and Madeline Carr. 2024. The geopolitics of technology standards: historical context for US EU and Chinese approaches. International Affairs 100 4 (July 2024) 1635\u20131652. https:\/\/doi.org\/10.1093\/ia\/iiae124","DOI":"10.1093\/ia\/iiae124"}],"event":{"name":"FAccT '25: The 2025 ACM Conference on Fairness, Accountability, and Transparency","acronym":"FAccT '25","location":"Athens Greece"},"container-title":["Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3715275.3732201","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,24]],"date-time":"2025-06-24T11:03:46Z","timestamp":1750763026000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3715275.3732201"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,23]]},"references-count":107,"alternative-id":["10.1145\/3715275.3732201","10.1145\/3715275"],"URL":"https:\/\/doi.org\/10.1145\/3715275.3732201","relation":{},"subject":[],"published":{"date-parts":[[2025,6,23]]},"assertion":[{"value":"2025-06-23","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}