{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T01:16:57Z","timestamp":1768007817327,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":102,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,6,3]],"date-time":"2024-06-03T00:00:00Z","timestamp":1717372800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["2131466\/2131477"],"award-info":[{"award-number":["2131466\/2131477"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,6,3]]},"DOI":"10.1145\/3630106.3659028","type":"proceedings-article","created":{"date-parts":[[2024,6,5]],"date-time":"2024-06-05T13:14:21Z","timestamp":1717593261000},"page":"2101-2112","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":4,"title":["Regulating Explainability in Machine Learning Applications -- Observations from a Policy Design Experiment"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4448-4804","authenticated-orcid":false,"given":"Nadia","family":"Nahar","sequence":"first","affiliation":[{"name":"Software and Societal Systems Department, School of Computer Science, Carnegie Mellon University, United States"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-5699-669X","authenticated-orcid":false,"given":"Jenny","family":"Rowlett","sequence":"additional","affiliation":[{"name":"Oberlin College, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-9096-301X","authenticated-orcid":false,"given":"Matthew","family":"Bray","sequence":"additional","affiliation":[{"name":"Yale University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-8247-5023","authenticated-orcid":false,"given":"Zahra Abba","family":"Omar","sequence":"additional","affiliation":[{"name":"Yale University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7711-3335","authenticated-orcid":false,"given":"Xenophon","family":"Papademetris","sequence":"additional","affiliation":[{"name":"Yale University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7420-751X","authenticated-orcid":false,"given":"Alka","family":"Menon","sequence":"additional","affiliation":[{"name":"Yale University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4450-4572","authenticated-orcid":false,"given":"Christian","family":"K\u00e4stner","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, United States"}]}],"member":"320","published-online":{"date-parts":[[2024,6,5]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems","author":"Alkhatib A.","year":"2019","unstructured":"[1] Alkhatib, A. and Bernstein, M. 2019. Street-Level Algorithms: A Theory at the Gaps Between Policy and Decisions. Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (2019), 1\u201313."},{"key":"e_1_3_2_1_2_1","volume-title":"Proceedings of the 17th Annual Computer Security Applications Conference","author":"Anderson R.","year":"2001","unstructured":"[2] Anderson, R. 2001. Why information security is hard - an economic perspective. Proceedings of the 17th Annual Computer Security Applications Conference (2001), 358\u2013365."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1147\/JRD.2019.2942288"},{"key":"e_1_3_2_1_4_1","volume-title":"Responsive Regulation: Transcending the Deregulation Debate","author":"Ayres I.","year":"1995","unstructured":"[4] Ayres, I. and Braithwaite, J. 1995. Responsive Regulation: Transcending the Deregulation Debate. Oxford University Press."},{"key":"e_1_3_2_1_5_1","volume-title":"Telematics and Informatics. 52","author":"Baik J.S.","year":"2020","unstructured":"[5] Baik, J.S. 2020. Data Privacy Against Innovation or Against Discrimination?: The Case of the California Consumer Privacy Act (CCPA). Telematics and Informatics. 52, (2020)."},{"key":"e_1_3_2_1_6_1","volume-title":"et al","author":"Baum K.","year":"2023","unstructured":"[6] Baum, K. et al. 2023. From fear to action: AI governance and opportunities for all. Frontiers in Computer Science. 5, (2023)."},{"key":"e_1_3_2_1_7_1","unstructured":"[7] Bender E. and Hannah A. 2023. AI causes real harm. Let\u2019s focus on that over the end-of-humanity hype. Scientific American."},{"key":"e_1_3_2_1_8_1","volume-title":"Race After Technology: Abolitionist Tools for the New Jim Code","author":"Benjamin R.","unstructured":"[8] Benjamin, R. 2019. Race After Technology: Abolitionist Tools for the New Jim Code. John Wiley & Sons."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3375624"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372860"},{"key":"e_1_3_2_1_11_1","unstructured":"[11] Blueprint for an AI Bill of Rights: 2022. https:\/\/www.whitehouse.gov\/ostp\/ai-bill-of-rights\/."},{"key":"e_1_3_2_1_12_1","unstructured":"[12] Broughel J. 2023. The Case For Artificial Intelligence Regulation Is Surprisingly Weak. Forbes Magazine."},{"key":"e_1_3_2_1_13_1","volume-title":"More than a Glitch: Confronting Race, Gender, and Ability Bias in Tech","author":"Broussard M.","unstructured":"[13] Broussard, M. 2023. More than a Glitch: Confronting Race, Gender, and Ability Bias in Tech. MIT Press."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"[14] Carpenter D. and Moss D.A. 2013. Preventing Regulatory Capture: Special Interest Influence and How to Limit it. Cambridge University Press.","DOI":"10.1017\/CBO9781139565875"},{"key":"e_1_3_2_1_15_1","unstructured":"[15] Claypoole T. and Dickinson W.B. 2023. Why We Shouldn\u2019t Talk About Regulating AI. Legaltech News."},{"key":"e_1_3_2_1_16_1","volume-title":"Is explainable artificial intelligence intrinsically valuable? AI & society. 37, 1","author":"Colaner N.","year":"2022","unstructured":"[16] Colaner, N. 2022. Is explainable artificial intelligence intrinsically valuable? AI & society. 37, 1 (2022), 231\u2013238."},{"key":"e_1_3_2_1_17_1","unstructured":"[17] Commission Federal Trade 2016. Using Consumer Reports for Credit Decisions: What to Know About Adverse Action and Risk-Based Pricing Notices. (2016)."},{"key":"e_1_3_2_1_18_1","first-page":"1","article-title":"Underspecification presents challenges for credibility in modern machine learning","volume":"23","author":"D\u2019Amour A.","year":"2022","unstructured":"[18] D\u2019Amour, A. et al. 2022. Underspecification presents challenges for credibility in modern machine learning. The Journal of Machine Learning Research. 23, 226 (2022), 1\u201361.","journal-title":"The Journal of Machine Learning Research."},{"key":"e_1_3_2_1_19_1","volume-title":"Insight - Amazon scraps secret AI recruiting tool that showed bias against women","author":"Dastin J.","unstructured":"[19] Dastin, J. 2018. Insight - Amazon scraps secret AI recruiting tool that showed bias against women. Reuters."},{"key":"e_1_3_2_1_20_1","volume-title":"et al","author":"De Paor A.","year":"2017","unstructured":"[20] De Paor, A. et al. 2017. Algorithmic governance: Developing a research agenda through the power of collective intelligence. Big Data and Society. 4, 2 (2017)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/2844110"},{"key":"e_1_3_2_1_22_1","volume-title":"Weak State: The Rights Revolution and the Rise of Human Resources Management Divisions. The American journal of sociology. 104, 2","author":"Dobbin F.","year":"1998","unstructured":"[22] Dobbin, F. and Sutton, J.R. 1998. The Strength of a Weak State: The Rights Revolution and the Rise of Human Resources Management Divisions. The American journal of sociology. 104, 2 (1998), 441\u2013476."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"crossref","unstructured":"[23] Edwards L. and Veale M. 2017. Slave to the algorithm? Why a\u2019right to an explanation\u2019is probably not the remedy you are looking for. Duke law and technology review. 16 (2017) 18.","DOI":"10.31228\/osf.io\/97upg"},{"key":"e_1_3_2_1_24_1","volume-title":"et al","author":"Ehsan U.","year":"2021","unstructured":"[24] Ehsan, U. et al. 2021. The Who in Explainable AI: How AI Background Shapes Perceptions of AI Explanations. arXiv [cs.HC]."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/2702123.2702556"},{"key":"e_1_3_2_1_26_1","volume-title":"France, Germany and Italy risk unravelling landmark AI Act negotiations","author":"EU","year":"2023","unstructured":"[26] EU: France, Germany and Italy risk unravelling landmark AI Act negotiations: 2023. https:\/\/www.amnesty.org\/en\/latest\/news\/2023\/11\/eu-france-germany-and-italy-risk-unravelling-landmark-ai-act-negotiations\/."},{"key":"e_1_3_2_1_27_1","unstructured":"[27] Executive Order on the Safe Secure and Trustworthy Development and Use of Artificial Intelligence: 2023. https:\/\/www.whitehouse.gov\/briefing-room\/presidential-actions\/2023\/10\/30\/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence\/."},{"key":"e_1_3_2_1_28_1","volume-title":"Proceedings of the 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Eykholt K.","year":"2018","unstructured":"[28] Eykholt, K. et al. 2018. Robust physical-world attacks on deep learning visual classification. Proceedings of the 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2018), 1625\u20131634."},{"key":"e_1_3_2_1_29_1","unstructured":"[29] Fafard Patrick Evidence and healthy public policy: Insights from health and political sciences: http:\/\/www.ncchpp.ca\/docs\/FafardEvidence08June.pdf."},{"key":"e_1_3_2_1_30_1","volume-title":"et al","author":"Feffer M.","year":"2024","unstructured":"[30] Feffer, M. et al. 2024. Red-Teaming for Generative AI: Silver Bullet or Security Theater? arXiv [cs.CY]."},{"key":"e_1_3_2_1_31_1","volume-title":"et al","author":"Ferreira G.","year":"2019","unstructured":"[31] Ferreira, G. et al. 2019. Design Dimensions for Software Certification: A Grounded Analysis. arXiv [cs.SE]."},{"key":"e_1_3_2_1_32_1","volume-title":"et al","author":"Fjeld J.","year":"2020","unstructured":"[32] Fjeld, J. et al. 2020. Principled Artificial Intelligence: Mapping Consensus in Ethical and Rights-Based Approaches to Principles for AI. Berkman Klein Center. 1, (2020)."},{"key":"e_1_3_2_1_33_1","first-page":"219","article-title":"Five Misunderstandings About Case-Study Research. Qualitative inquiry","volume":"12","author":"Flyvbjerg B.","year":"2006","unstructured":"[33] Flyvbjerg, B. 2006. Five Misunderstandings About Case-Study Research. Qualitative inquiry: QI. 12, 2 (2006), 219\u2013245.","journal-title":"QI."},{"key":"e_1_3_2_1_34_1","volume-title":"Proposed regulatory framework for modifications to Artificial Intelligence\/Machine Learning (AI\/ML)-based Software as a Medical Device (SaMD)","author":"Food and Drug Administration 2019.","year":"2019","unstructured":"[34] Food and Drug Administration 2019. Proposed regulatory framework for modifications to Artificial Intelligence\/Machine Learning (AI\/ML)-based Software as a Medical Device (SaMD). Department of Health and Human Services (United States). (2019)."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3458723"},{"key":"e_1_3_2_1_36_1","unstructured":"[36] Google Responsible Development of AI: https:\/\/ai.google\/static\/documents\/responsible-development-of-ai.pdf."},{"key":"e_1_3_2_1_37_1","volume-title":"Proceedings of the Hawaii International Conference on System Sciences (HICSS-52)","author":"Greene D.","year":"2019","unstructured":"[37] Greene, D. et al. 2019. Better, Nicer, Clearer, Fairer: A Critical Assessment of the Movement for Ethical Artificial Intelligence and Machine Learning. Proceedings of the Hawaii International Conference on System Sciences (HICSS-52) (2019)."},{"key":"e_1_3_2_1_38_1","volume-title":"et al","author":"Guha N.","year":"2023","unstructured":"[38] Guha, N. et al. 2023. Ai regulation has its own alignment problem: The technical and institutional feasibility of disclosure, registration, licensing, and auditing. George Washington Law Review, Forthcoming. (2023)."},{"key":"e_1_3_2_1_39_1","unstructured":"[39] Hagemann Ryan and Leclerc Jean-Marc Precision regulation for artificial intelligence: https:\/\/www.ibm.com\/policy\/wp-content\/uploads\/2023\/04\/IBM-AI-POV_FINAL2.pdf."},{"key":"e_1_3_2_1_40_1","volume-title":"Does the common criteria paradigm have a future? [security and privacy]","author":"Hearn J.","year":"2004","unstructured":"[40] Hearn, J. 2004. Does the common criteria paradigm have a future? [security and privacy]. IEEE security & privacy. 2, 1 (2004), 64\u201365."},{"key":"e_1_3_2_1_41_1","unstructured":"[41] Henshall W. 2023. E.U.\u2019s AI Regulation Could Be Softened After Pushback From Biggest Members. Time."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300809"},{"key":"e_1_3_2_1_43_1","volume-title":"FACT SHEET: Biden-Harris administration secures voluntary commitments from leading artificial intelligence companies to manage the risks posed by AI","author":"House W.","year":"2023","unstructured":"[43] House, W. 2023. FACT SHEET: Biden-Harris administration secures voluntary commitments from leading artificial intelligence companies to manage the risks posed by AI. The White House. (2023)."},{"key":"e_1_3_2_1_44_1","unstructured":"[44] Howlett M. and Ramesh M. 2003. Studying Public Policy: Policy Cycles and Policy Subsystems. Oxford University Press."},{"key":"e_1_3_2_1_45_1","volume-title":"et al","author":"Huang X.","year":"2020","unstructured":"[45] Huang, X. et al. 2020. A survey of safety and trustworthiness of deep neural networks: Verification, testing, adversarial attack and defence, and interpretability. Computer Science Review. 37, (2020), 100270."},{"key":"e_1_3_2_1_46_1","unstructured":"[46] Juelsen E. and Thoresen M.A. 2021. Shapley values in the context of GDPR: Can Shapley Values be used as a means of interpreting black-box machine learning models while also complying with the General Data Protection Regulation? (Master\u2019s thesis)."},{"key":"e_1_3_2_1_47_1","first-page":"1","article-title":"Design and Innovation in the Public Sector: Matters of Design in Policy-Making and Policy Implementation","volume":"1","author":"Junginger S.","year":"2013","unstructured":"[47] Junginger, S. 2013. Design and Innovation in the Public Sector: Matters of Design in Policy-Making and Policy Implementation. Annual Review of Policy Design. 1, 1 (2013), 1\u201311.","journal-title":"Annual Review of Policy Design."},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2012.29"},{"key":"e_1_3_2_1_49_1","unstructured":"[49] Kang C. 2023. OpenAI\u2019s Sam Altman Urges AI Regulation in Senate Hearing\u2019. The New York times."},{"key":"e_1_3_2_1_50_1","unstructured":"[50] K\u00e4stner C. 2022. Machine Learning in Production: From Models to Products."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"crossref","unstructured":"[51] Keblawi F. and Sullivan D. 2006. Applying the common criteria in systems engineering. IEEE security & privacy. 4 2 (2006) 50\u201355.","DOI":"10.1109\/MSP.2006.35"},{"key":"e_1_3_2_1_52_1","volume-title":"The Moral Limits of Predictive Practices: The Case of Credit-Based Insurance Scores. American sociological review. 84, 6","author":"Kiviat B.","year":"2019","unstructured":"[52] Kiviat, B. 2019. The Moral Limits of Predictive Practices: The Case of Credit-Based Insurance Scores. American sociological review. 84, 6 (2019), 1134\u20131158."},{"key":"e_1_3_2_1_53_1","volume-title":"To Regulate AI or Not? How should Governments React to the Artificial Intelligence Revolution?","author":"Krasadakis George","year":"2023","unstructured":"[53] Krasadakis, George, To Regulate AI or Not? How should Governments React to the Artificial Intelligence Revolution? 2023. https:\/\/medium.com\/60-leaders\/to-regulate-or-not-how-should-governments-react-to-the-ai-revolution-c254d176304f."},{"key":"e_1_3_2_1_54_1","volume-title":"Proceedings of the 37th International Conference on Machine Learning","author":"Kumar I.E.","year":"2020","unstructured":"[54] Kumar, I.E. et al. 2020. Problems with Shapley-value-based explanations as feature importance measures. Proceedings of the 37th International Conference on Machine Learning (2020), 5491\u20135500."},{"key":"e_1_3_2_1_55_1","volume-title":"et al","author":"Kurakin A.","year":"2016","unstructured":"[55] Kurakin, A. et al. 2016. Adversarial Machine Learning at Scale. arXiv [cs.CV]."},{"key":"e_1_3_2_1_56_1","first-page":"18","article-title":"Explainable AI","volume":"23","author":"Linardatos P.","year":"2020","unstructured":"[56] Linardatos, P. et al. 2020. Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy. 23, 1 (2020), 18.","journal-title":"A Review of Machine Learning Interpretability Methods. Entropy."},{"key":"e_1_3_2_1_57_1","volume-title":"Proc. IFIP-SEC. 91","author":"Lipner S.B.","year":"1991","unstructured":"[57] Lipner, S.B. 1991. Criteria, evaluation, and the international environment: where have we been, where are we going. Proc. IFIP-SEC. 91, (1991)."},{"key":"e_1_3_2_1_58_1","first-page":"3","article-title":"The Mythos of Model Interpretability: In machine learning, the concept of interpretability is both important and slippery","volume":"16","author":"Lipton Z.C.","year":"2018","unstructured":"[58] Lipton, Z.C. 2018. The Mythos of Model Interpretability: In machine learning, the concept of interpretability is both important and slippery. Queueing Systems. Theory and Applications. 16, 3 (Jun. 2018), 31\u201357.","journal-title":"Queueing Systems. Theory and Applications."},{"key":"e_1_3_2_1_59_1","volume-title":"Opt-Out Rights, Medical Information Usage, and Consumer Information Disposal. ISJLP. 2","author":"Liu K.K.","year":"2005","unstructured":"[59] Liu, K.K. 2005. Fair and Accurate Credit Transactions Act Regulations: Disclosure, Opt-Out Rights, Medical Information Usage, and Consumer Information Disposal. ISJLP. 2, (2005), 715."},{"key":"e_1_3_2_1_60_1","volume-title":"-I","author":"Lundberg S.M.","year":"2017","unstructured":"[60] Lundberg, S.M. and Lee, S.-I. 2017. A Unified Approach to Interpreting Model Predictions. Advances in Neural Information Processing Systems (NIPS). 30, (2017)."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3594064"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376445"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1109\/MC.2008.37"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"crossref","first-page":"61","DOI":"10.1017\/S0143814X00005006","article-title":"In and out of the revolving door: Making sense of regulatory capture","volume":"12","author":"Makkai T.","year":"1992","unstructured":"[64] Makkai, T. and Braithwaite, J. 1992. In and out of the revolving door: Making sense of regulatory capture. Journal of public policy. 12, 1 (1992), 61\u201378.","journal-title":"Journal of public policy."},{"key":"e_1_3_2_1_65_1","volume-title":"et al","author":"McGraw G.","year":"2020","unstructured":"[65] McGraw, G. et al. 2020. An architectural risk analysis of machine learning systems: Toward more secure machine learning. Technical report, Berryville Institute of Machine Learning. (2020)."},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-020-0214-1"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1353\/sor.2019.0022"},{"key":"e_1_3_2_1_68_1","volume-title":"Voluntary Commitments by Microsoft to Advance Responsible AI Innovation","author":"Microsoft","year":"2023","unstructured":"[68] Microsoft, Voluntary Commitments by Microsoft to Advance Responsible AI Innovation: 2023. https:\/\/blogs.microsoft.com\/wp-content\/uploads\/prod\/sites\/5\/2023\/07\/Microsoft-Voluntary-Commitments-July-21-2023.pdf."},{"key":"e_1_3_2_1_69_1","volume-title":"Proceedings of the Conference on Fairness, Accountability, and Transparency","author":"Mitchell M.","year":"2019","unstructured":"[69] Mitchell, M. et al. 2019. Model Cards for Model Reporting. Proceedings of the Conference on Fairness, Accountability, and Transparency (2019), 220\u2013229."},{"key":"e_1_3_2_1_70_1","volume-title":"Machine Learning","author":"Mitchell T.M.","unstructured":"[70] Mitchell, T.M. 1997. Machine Learning. McGraw-Hill."},{"key":"e_1_3_2_1_71_1","unstructured":"[71] Molnar C. 2020. Interpretable Machine Learning. Lulu.com."},{"key":"e_1_3_2_1_72_1","volume-title":"Nadia","author":"Nahar","year":"2024","unstructured":"[72] Nahar, Nadia, Supplementary Documents: Regulating Explainability in Machine Learning Applications: 2024. https:\/\/osf.io\/4xzpr\/."},{"key":"e_1_3_2_1_73_1","volume-title":"Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency","author":"Panigutti C.","year":"2023","unstructured":"[73] Panigutti, C. et al. 2023. The role of explainable AI in the context of the AI Act. Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency (2023), 1139\u20131150."},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1145\/3635715"},{"key":"e_1_3_2_1_75_1","volume-title":"Proceedings of the ACM on Human-Computer Interaction. 2, CSCW","author":"Passi S.","year":"2018","unstructured":"[75] Passi, S. and Jackson, S.J. 2018. Trust in Data Science: Collaboration, Translation, and Accountability in Corporate Data Science Projects. Proceedings of the ACM on Human-Computer Interaction. 2, CSCW (2018), 1\u201328."},{"key":"e_1_3_2_1_76_1","unstructured":"[76] People + AI Guidebook: https:\/\/pair.withgoogle.com\/guidebook\/."},{"key":"e_1_3_2_1_77_1","volume-title":"Exclusive: OpenAI Lobbied the EU to Water Down AI Regulation. Time.","author":"Perrigo B.","year":"2023","unstructured":"[77] Perrigo, B. 2023. Exclusive: OpenAI Lobbied the EU to Water Down AI Regulation. Time."},{"key":"e_1_3_2_1_78_1","volume-title":"The American political science review. 94, 2","author":"Pierson P.","year":"2000","unstructured":"[78] Pierson, P. 2000. Increasing Returns, Path Dependence, and the Study of Politics. The American political science review. 94, 2 (2000), 251\u2013267."},{"key":"e_1_3_2_1_79_1","volume-title":"Artificial Intelligence Act: MEPs adopt landmark law","author":"Press Releases","year":"2024","unstructured":"[79] Press Releases: Artificial Intelligence Act: MEPs adopt landmark law: 2024. https:\/\/www.europarl.europa.eu\/news\/en\/press-room\/20240308IPR19015\/artificial-intelligence-act-meps-adopt-landmark-law."},{"key":"e_1_3_2_1_80_1","volume-title":"et al","author":"Rong Y.","year":"2023","unstructured":"[80] Rong, Y. et al. 2023. Towards Human-Centered Explainable AI: A Survey of User Studies for Model Explanations. IEEE transactions on pattern analysis and machine intelligence. PP, (2023)."},{"key":"e_1_3_2_1_81_1","volume-title":"Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead. Nature machine intelligence. 1, 5","author":"Rudin C.","year":"2019","unstructured":"[81] Rudin, C. 2019. Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead. Nature machine intelligence. 1, 5 (2019), 206\u2013215."},{"key":"e_1_3_2_1_82_1","volume-title":"Proceedings of the 9th ACM international conference on Embedded software","author":"Rushby J.","year":"2011","unstructured":"[82] Rushby, J. 2011. New challenges in certification for aircraft software. Proceedings of the 9th ACM international conference on Embedded software (2011), 211\u2013218."},{"key":"e_1_3_2_1_83_1","volume-title":"The Action Research Guidebook: A Four-Stage Process for Educators and School Teams","author":"Sagor R.","unstructured":"[83] Sagor, R. 2011. The Action Research Guidebook: A Four-Stage Process for Educators and School Teams. Corwin Press."},{"key":"e_1_3_2_1_84_1","doi-asserted-by":"publisher","DOI":"10.1037\/0021-9010.88.3.444"},{"key":"e_1_3_2_1_85_1","volume-title":"Proceedings of the 3th International Conference on Quality of Information and Communications Technology","author":"Siebert J.","year":"2020","unstructured":"[85] Siebert, J. et al. 2020. Towards Guidelines for Assessing Qualities of Machine Learning Systems. Proceedings of the 3th International Conference on Quality of Information and Communications Technology (2020), 17\u201331."},{"key":"e_1_3_2_1_86_1","doi-asserted-by":"crossref","first-page":"6","DOI":"10.1177\/0002716213493066","article-title":"Organizational Challenges to Regulatory Enforcement and Compliance","volume":"649","author":"Silbey S.S.","year":"2013","unstructured":"[86] Silbey, S.S. 2013. Organizational Challenges to Regulatory Enforcement and Compliance. The Annals of the American Academy of Political and Social Science. 649, 1 (2013), 6\u201320.","journal-title":"The Annals of the American Academy of Political and Social Science."},{"key":"e_1_3_2_1_87_1","volume-title":"Beyond regulatory capture: Coproducing expertise for critical infrastructure protection. Regulation & governance. 12, 1","author":"Slayton R.","year":"2018","unstructured":"[87] Slayton, R. and Clark-Ginsberg, A. 2018. Beyond regulatory capture: Coproducing expertise for critical infrastructure protection. Regulation & governance. 12, 1 (2018), 115\u2013130."},{"key":"e_1_3_2_1_88_1","unstructured":"[88] Smiley L. 2023. The Legal Saga of Uber\u2019s Fatal Self-Driving Car Crash Is Over. Wired."},{"key":"e_1_3_2_1_89_1","doi-asserted-by":"crossref","unstructured":"[89] Sovrano F. and Vitali F. 2023. An objective metric for Explainable AI: How and why to estimate the degree of explainability. Knowledge-Based Systems. 278 (2023) 110866.","DOI":"10.1016\/j.knosys.2023.110866"},{"key":"e_1_3_2_1_90_1","volume-title":"et al","author":"Springer A.","year":"2018","unstructured":"[90] Springer, A. et al. 2018. Dice in the black box: User experiences with an inscrutable algorithm. arXiv [cs.HC]."},{"key":"e_1_3_2_1_91_1","unstructured":"[91] Stringer E.T. and Arag\u00f3n A.O. 2020. Action Research. SAGE Publications."},{"key":"e_1_3_2_1_92_1","doi-asserted-by":"publisher","DOI":"10.3390\/a16060271"},{"key":"e_1_3_2_1_93_1","doi-asserted-by":"publisher","DOI":"10.1108\/DPRG-01-2019-0007"},{"key":"e_1_3_2_1_94_1","doi-asserted-by":"publisher","DOI":"10.1109\/MC.2007.407"},{"key":"e_1_3_2_1_95_1","doi-asserted-by":"publisher","DOI":"10.1007\/s43681-022-00142-y"},{"key":"e_1_3_2_1_96_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3174014"},{"key":"e_1_3_2_1_97_1","unstructured":"[97] Vera Liao Q. and Varshney K.R. 2021. Human-Centered Explainable AI (XAI): From Algorithms to User Experiences. arXiv [cs.AI]."},{"key":"e_1_3_2_1_98_1","volume-title":"\u201cThe European Liability Regime For Large Language Models","author":"Volokh Eugene","year":"2023","unstructured":"[98] Volokh, Eugene, Journal of Free Speech Law: \u201cThe European Liability Regime For Large Language Models\u201d: 2023. https:\/\/reason.com\/volokh\/2023\/08\/11\/journal-of-free-speech-law-the-european-liability-regime-for-large-language-models\/."},{"key":"e_1_3_2_1_99_1","unstructured":"[99] Vranken Bram Big Tech lobbying is derailing the AI Act: https:\/\/www.socialeurope.eu\/big-tech-lobbying-is-derailing-the-ai-act."},{"key":"e_1_3_2_1_100_1","first-page":"76","article-title":"Why a right to explanation of automated decision-making does not exist in the general data protection regulation","volume":"7","author":"Wachter S.","year":"2017","unstructured":"[100] Wachter, S. et al. 2017. Why a right to explanation of automated decision-making does not exist in the general data protection regulation. SSRN Electronic Journal. 7, 2 (2017), 76\u201399.","journal-title":"SSRN Electronic Journal."},{"key":"e_1_3_2_1_101_1","doi-asserted-by":"crossref","unstructured":"[101] Wegrich K. and Jann W. 2007. Theories of the Policy Cycle. Handbook of Public Policy Analysis. Routledge. 43\u201362.","DOI":"10.1201\/9781420017007.pt2"},{"key":"e_1_3_2_1_102_1","volume-title":"The three challenges of AI regulation","author":"Wheeler Tom","year":"2023","unstructured":"[102] Wheeler, Tom, The three challenges of AI regulation: 2023. https:\/\/www.brookings.edu\/articles\/the-three-challenges-of-ai-regulation\/."}],"event":{"name":"FAccT '24: The 2024 ACM Conference on Fairness, Accountability, and Transparency","location":"Rio de Janeiro Brazil","acronym":"FAccT '24"},"container-title":["The 2024 ACM Conference on Fairness, Accountability, and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3630106.3659028","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3630106.3659028","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T23:57:07Z","timestamp":1750291027000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3630106.3659028"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,3]]},"references-count":102,"alternative-id":["10.1145\/3630106.3659028","10.1145\/3630106"],"URL":"https:\/\/doi.org\/10.1145\/3630106.3659028","relation":{},"subject":[],"published":{"date-parts":[[2024,6,3]]},"assertion":[{"value":"2024-06-05","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}