{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T17:28:27Z","timestamp":1775150907249,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":58,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,6,23]]},"DOI":"10.1145\/3715275.3732080","type":"proceedings-article","created":{"date-parts":[[2025,6,23]],"date-time":"2025-06-23T17:03:13Z","timestamp":1750698193000},"page":"1211-1223","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Promising Topics for US\u2013China Dialogues on AI Risks and Governance"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-6378-5086","authenticated-orcid":false,"given":"Saad","family":"Siddiqui","sequence":"first","affiliation":[{"name":"Safe AI Forum, London, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0395-784X","authenticated-orcid":false,"given":"Lujain","family":"Ibrahim","sequence":"additional","affiliation":[{"name":"University of Oxford, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-5031-5775","authenticated-orcid":false,"given":"Kristy","family":"Loke","sequence":"additional","affiliation":[{"name":"Independent, Ottawa, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4006-5426","authenticated-orcid":false,"given":"Stephen","family":"Clare","sequence":"additional","affiliation":[{"name":"Centre for the Governance of AI, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-8502-377X","authenticated-orcid":false,"given":"Marianne","family":"Lu","sequence":"additional","affiliation":[{"name":"Stanford University, Stanford, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-3380-9304","authenticated-orcid":false,"given":"Aris","family":"Richardson","sequence":"additional","affiliation":[{"name":"Centre for the Governance of AI, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4602-6527","authenticated-orcid":false,"given":"Conor","family":"McGlynn","sequence":"additional","affiliation":[{"name":"Harvard University, Boston, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0327-8193","authenticated-orcid":false,"given":"Jeffrey","family":"Ding","sequence":"additional","affiliation":[{"name":"George Washington University, Washington, D.C., USA"}]}],"member":"320","published-online":{"date-parts":[[2025,6,23]]},"reference":[{"key":"e_1_3_3_2_2_2","unstructured":"2023. Executive Order 14105: Addressing United States Investments in Certain National Security Technologies and Products in Countries of Concern. Federal Register 88 154 (2023) 54867\u201354872. https:\/\/www.federalregister.gov\/documents\/2023\/11\/01\/2023-24283\/safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence"},{"key":"e_1_3_3_2_3_2","unstructured":"2023. Executive Order 14110: Safe Secure and Trustworthy Development and Use of Artificial Intelligence. Federal Register 88 210 (2023)."},{"key":"e_1_3_3_2_4_2","volume-title":"White Paper on the Governance and Use of Generative Artificial Intelligence","author":"Center Alibaba AI Governance Research","year":"2023","unstructured":"Alibaba AI Governance Research Center. 2023. White Paper on the Governance and Use of Generative Artificial Intelligence. White Paper. Alibaba AI Governance Research Center. https:\/\/mp.weixin.qq.com\/mp\/appmsgalbum?__biz=Mzg4MTY2MzUyNA==&action=getalbum&album_id=3187743423251611652&scene=173&from_msgid=2247572341&from_itemidx=1&count=3&nolastread=1#wechat_redirect"},{"key":"e_1_3_3_2_5_2","unstructured":"Oxford Analytica. 2023. UK AI Summit will promote some global cooperation. Emerald Expert Briefingsoxan-db (2023)."},{"key":"e_1_3_3_2_6_2","volume-title":"The Claude 3 Model Family: Opus, Sonnet, Haiku","year":"2024","unstructured":"Anthropic. 2024. The Claude 3 Model Family: Opus, Sonnet, Haiku. Model Card. Anthropic. https:\/\/www-cdn.anthropic.com\/f2986af8d052f26236f6251da62d16172cfabd6e\/claude-3-model-card.pdf"},{"key":"e_1_3_3_2_7_2","volume-title":"Keep your AI claims in check","author":"Atleson Michael","year":"2023","unstructured":"Michael Atleson. 2023. Keep your AI claims in check. Federal Trade Commission. https:\/\/www.ftc.gov\/business-guidance\/blog\/2023\/02\/keep-your-ai-claims-check"},{"key":"e_1_3_3_2_8_2","volume-title":"Baichuan 2: Open Large-scale Language Models","author":"Inc. Baichuan","year":"2023","unstructured":"Baichuan Inc.2023. Baichuan 2: Open Large-scale Language Models. Technical Report. Baichuan Inc.https:\/\/cdn.baichuan-ai.com\/paper\/Baichuan2-technical-report.pdf"},{"key":"e_1_3_3_2_9_2","doi-asserted-by":"publisher","unstructured":"Yoshua Bengio et\u00a0al. 2024. Managing Extreme AI Risks Amid Rapid Progress. Science 384 (2024) 842\u2013845. https:\/\/doi.org\/10.1126\/science.adn0117","DOI":"10.1126\/science.adn0117"},{"key":"e_1_3_3_2_10_2","unstructured":"Joseph\u00a0R Biden. 2023. Executive order on the safe secure and trustworthy development and use of artificial intelligence. (2023)."},{"key":"e_1_3_3_2_11_2","volume-title":"Key Takeaways from the AI Seoul Summit 2024","author":"Birch Jessica","year":"2024","unstructured":"Jessica Birch and \u00d6yk\u00fc \u00d6zf\u0131rat. 2024. Key Takeaways from the AI Seoul Summit 2024. Access Partnership. https:\/\/accesspartnership.com\/key-takeaways-from-the-ai-seoul-summit-2024\/"},{"key":"e_1_3_3_2_12_2","unstructured":"Tom Bristow. 2025. Britain dances to JD Vance\u2019s tune as it renames AI institute. Politico EU. https:\/\/www.politico.eu\/article\/jd-vance-britain-ai-safety-institute-aisi-security\/"},{"key":"e_1_3_3_2_13_2","doi-asserted-by":"crossref","unstructured":"Kolja Brockmann. 2022. Applying Export Controls to AI: Current Coverage and Potential Future Controls. Armament Arms Control and Artificial Intelligence: The Janus-faced Nature of Machine Learning in the Military Realm (2022) 193\u2013209.","DOI":"10.1007\/978-3-031-11043-6_14"},{"key":"e_1_3_3_2_14_2","volume-title":"Authoritative large model AI Safety Benchmark first round results officially released","author":"Technology China Academy of Information and Communications","year":"2024","unstructured":"China Academy of Information and Communications Technology. 2024. Authoritative large model AI Safety Benchmark first round results officially released. China Academy of Information and Communications Technology. https:\/\/mp.weixin.qq.com\/s\/3FcLBHCy_oVaaj-2Ca9zag Translated by Jeffrey Ding, April 2024."},{"key":"e_1_3_3_2_15_2","volume-title":"Registration is now open: The first batch of evaluations of intelligent applications based on large models has been officially launched, and the standard publicity and implementation meeting has been successfully held","author":"Technology China Academy of Information and Communications","year":"2024","unstructured":"China Academy of Information and Communications Technology. 2024. Registration is now open: The first batch of evaluations of intelligent applications based on large models has been officially launched, and the standard publicity and implementation meeting has been successfully held. China Academy of Information and Communications Technology. https:\/\/mp.weixin.qq.com\/s\/V-LhB23hEx1n_tCxrvKSUA"},{"key":"e_1_3_3_2_16_2","unstructured":"China Law Society. [n. d.]. Scholars\u2019 Draft Law on AI."},{"key":"e_1_3_3_2_17_2","doi-asserted-by":"publisher","DOI":"10.5281\/zenodo.10974163"},{"key":"e_1_3_3_2_18_2","unstructured":"Jon Chun Christian Schroeder de Witt and Katherine Elkins. 2024. Comparative Global AI Regulation: Policy Perspectives from the EU China and the US. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2410.21279 (10 2024). https:\/\/arxiv.org\/pdf\/2410.21279"},{"key":"e_1_3_3_2_19_2","unstructured":"Concordia. 2024. China\u2019s AI Safety Evaluations Ecosystem. AI Safety in China (9 2024). https:\/\/aisafetychina.substack.com\/p\/chinas-ai-safety-evaluations-ecosystem"},{"key":"e_1_3_3_2_20_2","volume-title":"Provisions on the Management of Algorithmic Recommendations in Internet Information Services","author":"China Cyberspace Administration of","year":"2021","unstructured":"Cyberspace Administration of China. 2021. Provisions on the Management of Algorithmic Recommendations in Internet Information Services. Regulation. Cyberspace Administration of China. http:\/\/www.cac.gov.cn\/2022-01\/04\/c_1642894606364259.htm Translated by China Law Translate, January 2022."},{"key":"e_1_3_3_2_21_2","unstructured":"Cyberspace Administration of China. 2022. Provisions on the Administration of Deep Synthesis Internet Information Services."},{"key":"e_1_3_3_2_22_2","unstructured":"Cyberspace Administration of China. 2023. Interim Measures for the Management of Generative Artificial Intelligence Services. http:\/\/www.cac.gov.cn\/2023-07\/13\/c_1690898327029107.htm"},{"key":"e_1_3_3_2_23_2","unstructured":"Bob Davis. 2024. Back on Track? The Wire China (1 2024). https:\/\/www.thewirechina.com\/2024\/01\/14\/back-on-track-two-dialogues-u-s-china\/"},{"key":"e_1_3_3_2_24_2","unstructured":"Jeffrey Ding. 2024. ChinAI #261: First results from CAICT\u2019s AI Safety Benchmark. ChinAI Newsletter (4 2024). https:\/\/chinai.substack.com\/p\/chinai-261-first-results-from-caicts"},{"key":"e_1_3_3_2_25_2","doi-asserted-by":"crossref","unstructured":"Jeffrey Ding. 2024. Keep your enemies safer: technical cooperation and transferring nuclear safety and security technologies. European Journal of International Relations (2024) 13540661241246622.","DOI":"10.1177\/13540661241246622"},{"key":"e_1_3_3_2_26_2","volume-title":"A Growing Yard: The Biden Administration\u2019s China Export Controls Are Ensnaring CPUs","author":"Feldgoise Jacob","year":"2024","unstructured":"Jacob Feldgoise, Hanna Dohmen, and Brian Love. 2024. A Growing Yard: The Biden Administration\u2019s China Export Controls Are Ensnaring CPUs. Technical Report. Center for Security and Emerging Technology."},{"key":"e_1_3_3_2_27_2","unstructured":"Gemini Team Google. 2023. Gemini: A Family of Highly Capable Multimodal Models. (12 2023). arxiv:https:\/\/arXiv.org\/abs\/2312.11805\u00a0[cs.CL]"},{"key":"e_1_3_3_2_28_2","volume-title":"The Anatomy of AI Rules: A Systematic Comparison of AI Rules Across the Globe","author":"Giardini Tommaso","year":"2024","unstructured":"Tommaso Giardini and Johannes Fritz. 2024. The Anatomy of AI Rules: A Systematic Comparison of AI Rules Across the Globe. Technical Report. Digital Policy Alert. https:\/\/digitalpolicyalert.org\/ai-rules\/the-anatomy-of-AI-rules"},{"key":"e_1_3_3_2_29_2","volume-title":"Governing Silicon Valley and Shenzhen: Assessing a New Era of Artificial Intelligence Governance in the US and China","author":"Hine Emmie","year":"2023","unstructured":"Emmie Hine. 2023. Governing Silicon Valley and Shenzhen: Assessing a New Era of Artificial Intelligence Governance in the US and China. Technical Report. https:\/\/ssrn.com\/abstract=4553087"},{"key":"e_1_3_3_2_30_2","volume-title":"Artificial Intelligence with American Values and Chinese Characteristics: A Comparative Analysis of American and Chinese Governmental AI Policies","author":"Hine Emmie","year":"2022","unstructured":"Emmie Hine and Luciano Floridi. 2022. Artificial Intelligence with American Values and Chinese Characteristics: A Comparative Analysis of American and Chinese Governmental AI Policies. Technical Report. https:\/\/ssrn.com\/abstract=4006332"},{"key":"e_1_3_3_2_31_2","unstructured":"Michael\u00a0C Horowitz. 2018. Artificial intelligence international competition and the balance of power. 2018 22 (2018)."},{"key":"e_1_3_3_2_32_2","unstructured":"Geneva\u00a0Graduate institute. 2024. United Nations AI resolution: A significant global policy effort to harness the technology for Sustainable Development: Iheid FC. https:\/\/executive.graduateinstitute.ch\/communications\/news\/united-nations-ai-resolution-significant-global-policy-effort-harness#:\u00a0:text=The%20unanimous%20adoption%20of%20the and%20sustainable%20use%20of%20AI."},{"key":"e_1_3_3_2_33_2","volume-title":"Software Must Be Secure by Design, and Artificial Intelligence Is No Exception","author":"Lai Christine","year":"2023","unstructured":"Christine Lai and Jonathan Spring. 2023. Software Must Be Secure by Design, and Artificial Intelligence Is No Exception. Cybersecurity and Infrastructure Security Agency. https:\/\/www.cisa.gov\/news-events\/news\/software-must-be-secure-design-and-artificial-intelligence-no-exception"},{"key":"e_1_3_3_2_34_2","volume-title":"Guidelines for the Construction of a National Comprehensive Standardization System for the Artificial Intelligence Industry","author":"Technology Ministry of Industry and Information","year":"2024","unstructured":"Ministry of Industry and Information Technology. 2024. Guidelines for the Construction of a National Comprehensive Standardization System for the Artificial Intelligence Industry. Guidelines. Ministry of Industry and Information Technology. https:\/\/www.miit.gov.cn\/gzcy\/yjzj\/art\/2024\/art_983199be076649d494690135c0b4d168.html"},{"key":"e_1_3_3_2_35_2","volume-title":"U.S. Leadership in AI: A Plan for Federal Engagement in Developing Technical Standards and Related Tools","author":"Technology National Institute of Standards and","year":"2019","unstructured":"National Institute of Standards and Technology. 2019. U.S. Leadership in AI: A Plan for Federal Engagement in Developing Technical Standards and Related Tools. Technical Report. National Institute of Standards and Technology. https:\/\/www.nist.gov\/system\/files\/documents\/2019\/08\/10\/ai_standards_fedengagement_plan_9aug2019.pdf"},{"key":"e_1_3_3_2_36_2","volume-title":"Artificial Intelligence Risk Management Framework (AI RMF 1.0)","author":"Technology National Institute of Standards and","year":"2023","unstructured":"National Institute of Standards and Technology. 2023. Artificial Intelligence Risk Management Framework (AI RMF 1.0). Technical Report. U.S. Department of Commerce. https:\/\/nvlpubs.nist.gov\/nistpubs\/ai\/NIST.AI.100-1.pdf"},{"key":"e_1_3_3_2_37_2","volume-title":"White Paper on AI Security Evaluation","author":"Products National Quality Supervision and Inspection Center for Speech and Image Recognition","year":"2023","unstructured":"National Quality Supervision and Inspection Center for Speech and Image Recognition Products. 2023. White Paper on AI Security Evaluation. White Paper. National Quality Supervision and Inspection Center for Speech and Image Recognition Products."},{"key":"e_1_3_3_2_38_2","volume-title":"White Paper on AI Security Evaluation","author":"Products National Quality Supervision and Inspection Center for Speech and Image Recognition","year":"2021","unstructured":"National Quality Supervision and Inspection Center for Speech and Image Recognition Products, National Industrial Information Security Development Research Center, and Institute of Artificial Intelligence. 2021. White Paper on AI Security Evaluation. White Paper. National Quality Supervision and Inspection Center for Speech and Image Recognition Products. https:\/\/www.realai.ai\/media\/upload\/AI-research\/AI-results\/2021%EF%BC%89.pdf"},{"key":"e_1_3_3_2_39_2","volume-title":"Artificial Intelligence Computing Platform Information Security Framework","author":"China National Technical Committee 260 on Cybersecurity of Standardization Administration of","year":"2023","unstructured":"National Technical Committee 260 on Cybersecurity of Standardization Administration of China. 2023. Artificial Intelligence Computing Platform Information Security Framework. Technical Standard. https:\/\/www.tc260.org.cn\/front\/bzzqyjDetail.html?id=20230515154409898112&norm_id=20221102142806&recode_id=51281"},{"key":"e_1_3_3_2_40_2","volume-title":"Basic Safety Requirements for Generative Artificial Intelligence Services","author":"China National Technical Committee 260 on Cybersecurity of Standardization Administration of","year":"2024","unstructured":"National Technical Committee 260 on Cybersecurity of Standardization Administration of China. 2024. Basic Safety Requirements for Generative Artificial Intelligence Services. Technical Standard. https:\/\/perma.cc\/GU3Q-GAJ3 Translated by the Center for Security and Emerging Technology, April 2024."},{"key":"e_1_3_3_2_41_2","volume-title":"Standards Regarding Security Requirements for Automated Decision-Making Based on Personal Information","author":"China National Technical Committee 260 on Cybersecurity of Standardization Administration of","year":"2024","unstructured":"National Technical Committee 260 on Cybersecurity of Standardization Administration of China. 2024. Standards Regarding Security Requirements for Automated Decision-Making Based on Personal Information. Technical Standard."},{"key":"e_1_3_3_2_42_2","unstructured":"OpenAI. 2023. GPT-4 Technical Report. (3 2023). arxiv:https:\/\/arXiv.org\/abs\/2303.08774\u00a0[cs.CL]"},{"key":"e_1_3_3_2_43_2","unstructured":"Hadrien Pouget Claire Dennis J Batemen Robert\u00a0F Trager Renan Araujo Haydn Belfield Belinda Cleeland Malou Estier Gideon Futerman Oliver Guest et\u00a0al. 2024. The Future of International Scientific Assessments of AI\u2019s Risks."},{"key":"e_1_3_3_2_44_2","unstructured":"Jarrett Renshaw and Trevor Hunnicutt. 2024. Biden Xi agree that humans not AI should control nuclear arms. Reuters. https:\/\/www.reuters.com\/world\/biden-xi-agreed-that-humans-not-ai-should-control-nuclear-weapons-white-house-2024-11-16\/"},{"key":"e_1_3_3_2_45_2","unstructured":"Jarrett Renshaw and Trevor Hunnicutt. 2024. Biden Xi agree that humans not AI should control nuclear arms. Reuters (11 2024). https:\/\/www.reuters.com\/world\/biden-xi-agreed-that-humans-not-ai-should-control-nuclear-weapons-white-house-2024-11-16\/"},{"key":"e_1_3_3_2_46_2","unstructured":"John Revill Emma Farge and David Brunnstrom. 2025. Trump hails China talks says two sides negotiated \u2019total reset\u2019 in Geneva. Reuters. https:\/\/www.reuters.com\/world\/china\/china-us-trade-talks-begin-geneva-2025-05-10\/"},{"key":"e_1_3_3_2_47_2","volume-title":"AI Governance Whitepaper","year":"2022","unstructured":"SenseTime. 2022. AI Governance Whitepaper. White Paper. SenseTime. https:\/\/oss.sensetime.com\/20221014\/3ec8d5aaf00d134573f1e0d214ad9270\/AI%20Governance%20for%20Balanced%20Development%20White%20Paper-eng-final.pdf"},{"key":"e_1_3_3_2_48_2","volume-title":"AI (and other) Companies: Quietly Changing Your Terms of Service Could Be Unfair or Deceptive","author":"Protection Staff in the Office of Technology and The Division of Privacy and Identity","year":"2024","unstructured":"Staff in the Office of Technology and The Division of Privacy and Identity Protection. 2024. AI (and other) Companies: Quietly Changing Your Terms of Service Could Be Unfair or Deceptive. Federal Trade Commission. https:\/\/www.ftc.gov\/policy\/advocacy-research\/tech-at-ftc\/2024\/02\/ai-other-companies-quietly-changing-your-terms-service-could-be-unfair-or-deceptive"},{"key":"e_1_3_3_2_49_2","volume-title":"Information Security Technology \u2013 Assessment Specification for Security of Machine Learning Algorithms","author":"China Standardization Administration of","unstructured":"Standardization Administration of China. [n. d.]. Information Security Technology \u2013 Assessment Specification for Security of Machine Learning Algorithms. Technical Standard. Standardization Administration of China."},{"key":"e_1_3_3_2_50_2","volume-title":"Tencent Large Model Security and Safety Report","author":"Institute Tencent Research","year":"2024","unstructured":"Tencent Research Institute, Tencent Zhuque Lab, Tencent Hunyuan Model Team, Tsinghua Shenzhen International Graduate School, and Zhejiang University State Key Lab of Blockchain and Data Security. 2024. Tencent Large Model Security and Safety Report. Technical Report. Tencent Research Institute. https:\/\/mp.weixin.qq.com\/s\/KCWw9gBwUnzywyNW_K8-4A Translated by Jeffrey Ding, January 2024."},{"key":"e_1_3_3_2_51_2","unstructured":"The White House. 2023. FACT SHEET: Biden-Harris Administration Secures Voluntary Commitments from Leading Artificial Intelligence Companies to Manage the Risks Posed by AI. https:\/\/www.whitehouse.gov\/briefing-room\/statements-releases\/2023\/07\/21\/fact-sheet-biden-harris-administration-secures-voluntary-commitments-from-leading-artificial-intelligence-companies-to-manage-the-risks-posed-by-ai"},{"key":"e_1_3_3_2_52_2","unstructured":"The White House. 2025. Removing Barriers to American Leadership in Artificial Intelligence. https:\/\/www.whitehouse.gov\/presidential-actions\/2025\/01\/removing-barriers-to-american-leadership-in-artificial-intelligence\/"},{"key":"e_1_3_3_2_53_2","unstructured":"United Kingdom Government. 2023. The Bletchley Declaration by Countries Attending the AI Safety Summit 1-2 November 2023. https:\/\/www.gov.uk\/government\/publications\/ai-safety-summit-2023-the-bletchley-declaration\/the-bletchley-declaration-by-countries-attending-the-ai-safety-summit-1-2-november-2023 Accessed: 2025-01-17."},{"key":"e_1_3_3_2_54_2","volume-title":"Select Issues: Assessing Adverse Impact in Software, Algorithms, and Artificial Intelligence Used in Employment Selection Procedures Under Title VII of the Civil Rights Act of 1964","author":"Commission U.S. Equal Employment Opportunity","year":"2023","unstructured":"U.S. Equal Employment Opportunity Commission. 2023. Select Issues: Assessing Adverse Impact in Software, Algorithms, and Artificial Intelligence Used in Employment Selection Procedures Under Title VII of the Civil Rights Act of 1964. Guidance Title VII, 29 CFR Part 1607. U.S. Equal Employment Opportunity Commission. https:\/\/www.eeoc.gov\/laws\/guidance\/select-issues-assessing-adverse-impact-software-algorithms-and-artificial"},{"key":"e_1_3_3_2_55_2","unstructured":"Nicholas Welch. 2024. Tech policy at the Third plenum. https:\/\/www.chinatalk.media\/p\/tech-policy-at-the-third-plenum"},{"key":"e_1_3_3_2_56_2","doi-asserted-by":"crossref","unstructured":"Xiangning Wu. 2020. Technology power and uncontrolled great power strategic competition between China and the United States. China International Strategy Review 2 1 (2020) 99\u2013119.","DOI":"10.1007\/s42533-020-00040-0"},{"key":"e_1_3_3_2_57_2","unstructured":"Aohan Zeng et\u00a0al. 2023. GLM-130B: An Open Bilingual pre-Trained Model. (10 2023). arxiv:https:\/\/arXiv.org\/abs\/2210.02414\u00a0[cs.CL]"},{"key":"e_1_3_3_2_58_2","doi-asserted-by":"crossref","unstructured":"Yi Zeng Kevin Klyman Andy Zhou Yu Yang Minzhou Pan Ruoxi Jia Dawn Song Percy Liang and Bo Li. 2024. AI Risk Categorization Decoded (AIR 2024): From Government Regulations to Corporate Policies. (6 2024). arxiv:https:\/\/arXiv.org\/abs\/2406.17864\u00a0[cs.AI]","DOI":"10.70777\/si.v1i1.10603"},{"key":"e_1_3_3_2_59_2","unstructured":"Lim\u00a0Min Zhang. 2024. US China agree to expand military talks continue AI cooperation after Sullivan-Wang meet. The Straits Times (8 2024). https:\/\/www.straitstimes.com\/asia\/east-asia\/us-china-agree-to-expand-military-talks-continue-ai-cooperation-after-sullivan-wang-meet"}],"event":{"name":"FAccT '25: The 2025 ACM Conference on Fairness, Accountability, and Transparency","location":"Athens Greece","acronym":"FAccT '25"},"container-title":["Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3715275.3732080","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,24]],"date-time":"2025-06-24T11:24:08Z","timestamp":1750764248000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3715275.3732080"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,23]]},"references-count":58,"alternative-id":["10.1145\/3715275.3732080","10.1145\/3715275"],"URL":"https:\/\/doi.org\/10.1145\/3715275.3732080","relation":{},"subject":[],"published":{"date-parts":[[2025,6,23]]},"assertion":[{"value":"2025-06-23","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}