{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T06:26:02Z","timestamp":1771827962425,"version":"3.50.1"},"reference-count":68,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2025,8,1]],"date-time":"2025-08-01T00:00:00Z","timestamp":1754006400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,8,1]],"date-time":"2025-08-01T00:00:00Z","timestamp":1754006400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Digit. Soc."],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s44206-025-00221-7","type":"journal-article","created":{"date-parts":[[2025,8,5]],"date-time":"2025-08-05T20:41:04Z","timestamp":1754426464000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Design versus reality: assessing the results and compliance of algorithmic impact assessments"],"prefix":"10.1007","volume":"4","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1672-5244","authenticated-orcid":false,"given":"Ana","family":"Brandusescu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0720-3498","authenticated-orcid":false,"given":"Ren\u00e9e E.","family":"Sieber","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,5]]},"reference":[{"key":"221_CR1","unstructured":"Ada Lovelace Institute. (2022). AIA User Guide. https:\/\/www.adalovelaceinstitute.org\/resource\/aia-user-guide\/."},{"key":"221_CR2","doi-asserted-by":"crossref","unstructured":"Ananny, M., Crawford, K. (2018). Seeing without knowing: Limitations of the transparency ideal and its application to algorithmic accountability. New Media and Society, 20(3), 973\u2013989.","DOI":"10.1177\/1461444816676645"},{"key":"221_CR3","doi-asserted-by":"publisher","unstructured":"Ashar, A., Ginena, K., Cipollone, M., Barreto, R., & Cramer, H. (2024). Algorithmic impact assessments at scale: Practitioners\u2019 challenges and needs. Journal of Online Trust and Safety, 2(4). https:\/\/doi.org\/10.54501\/jots.v2i4.206.","DOI":"10.54501\/jots.v2i4.206"},{"key":"221_CR4","doi-asserted-by":"crossref","unstructured":"Attard-Frost, B. (2023). Generative AI systems: Impacts on artists & creators and related gaps in the artificial intelligence and data act. Submission to the Standing committee on industry and technology. https:\/\/www.ourcommons.ca\/Content\/Committee\/441\/INDU\/Brief\/BR12541028\/br-external\/AttardFrostBlair-e.pdf.","DOI":"10.2139\/ssrn.4468637"},{"issue":"2","key":"221_CR5","doi-asserted-by":"publisher","first-page":"101929","DOI":"10.1016\/j.giq.2024.101929","volume":"41","author":"B. Attard-Frost","year":"2024","unstructured":"Attard-Frost, B., Brandusescu, A., & Lyons, K. (2024). The governance of artificial intelligence in Canada: Findings and opportunities from a review of 84 AI governance initiatives. Government Information Quarterly, 41(2), 101929. https:\/\/doi.org\/10.1016\/j.giq.2024.101929.","journal-title":"Government Information Quarterly"},{"issue":"4","key":"221_CR6","doi-asserted-by":"publisher","first-page":"447","DOI":"10.1111\/j.1468-0386.2007.00378.x","volume":"13","author":"M. Bovens","year":"2007","unstructured":"Bovens, M. (2007). Analysing and assessing accountability: A conceptual framework. European Law Journal, 13(4), 447\u2013468. https:\/\/doi.org\/10.1111\/j.1468-0386.2007.00378.x.","journal-title":"European Law Journal"},{"key":"221_CR7","doi-asserted-by":"publisher","first-page":"182","DOI":"10.1093\/oxfordhb\/9780199226443.003.0009","volume-title":"The Oxford handbook of public management","author":"M. Bovens","year":"2009","unstructured":"Bovens, M. (2009). Public accountability. In E. Ferlie, L. L.E Jr, & C. Pollitt (Eds.), The Oxford handbook of public management (pp. 182\u2013208). Oxford University Press. https:\/\/doi.org\/10.1093\/oxfordhb\/9780199226443.003.0009."},{"key":"221_CR8","volume-title":"Centre for interdisciplinary research on montreal","author":"A. Brandusescu","year":"2021","unstructured":"Brandusescu, A. (2021). Artificial intelligence policy and funding in Canada: Public investments, private interests. In Centre for interdisciplinary research on montreal. Montreal, QC: McGill University. https:\/\/www.mcgill.ca\/centre-montreal\/files\/centremontreal\/aipolicyandfunding_report_updated_mar5.pdf"},{"key":"221_CR9","doi-asserted-by":"publisher","first-page":"e40","DOI":"10.1017\/dap.2025.17","volume":"7","author":"A. Brandusescu","year":"2025","unstructured":"Brandusescu, A., & Sieber, R. E. (2025). Missed opportunities in AI regulation: Lessons from Canada\u2019s AI and data act. Data & Policy, 7, e40. https:\/\/doi.org\/10.1017\/dap.2025.17.","journal-title":"Data & Policy"},{"issue":"2133","key":"221_CR10","doi-asserted-by":"publisher","first-page":"20180080","DOI":"10.1098\/rsta.2018.0080","volume":"376","author":"C. Cath","year":"2018","unstructured":"Cath, C. (2018). Governing artificial intelligence: Ethical, legal and technical opportunities and challenges. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 376(2133), 20180080. https:\/\/doi.org\/10.1098\/rsta.2018.0080.","journal-title":"Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences"},{"key":"221_CR11","unstructured":"Centre, S. (2024). Tracking automated government \u2018TAG\u2019 register Canada. Western university. http:\/\/tagregistercanada.ca\/."},{"key":"221_CR12","unstructured":"Christian, G. (2024). CBSA border surveillance: The dangerous expansion of facial recognition technology. Canadian immigration lawyers association. https:\/\/cila.co\/cbsa-border-surveillance-the-dangerous-expansion-of-facial-recognition-technology\/."},{"key":"221_CR13","unstructured":"Council of Europe. (2021). Human rights, democracy and rule of law impact assessment of AI systems. ad hoc committee on artificial intelligence (cahai) policy development group (CAHAI-PDG). https:\/\/rm.coe.int\/cahai-pdg-2021-02-subworkinggroup1-ai-impact-assessment-v1-2769-4229-7\/1680a1bd2d."},{"key":"221_CR14","unstructured":"Curry, B. (2024). Senior officials reject allegation that employee linked to ArriveCan was told to lie to investigators. The globe and mail. https:\/\/www.theglobeandmail.com\/politics\/article-arrivecan-employees-allegations-investigation."},{"issue":"1","key":"221_CR15","doi-asserted-by":"publisher","first-page":"74","DOI":"10.7202\/1106045ar","volume":"20","author":"P. Daly","year":"2023","unstructured":"Daly, P. (2023). Mapping artificial intelligence use in the government of Canada. Revue Gouvernance\/Governance Review, 20(1), 74\u201395. https:\/\/www.erudit.org\/en\/journals\/gouvernance\/2023-v20-n1-gouvernance08729\/1106045ar\/.","journal-title":"Revue Gouvernance\/Governance Review"},{"key":"221_CR16","unstructured":"Darbyshire, T. (2022). In praise of the Canadian algorithmic impact assessment framework. Tech UK. https:\/\/www.techuk.org\/resource\/in-praise-of-the-canadian-algorithmic-impact-assessment-framework.html."},{"key":"221_CR17","volume-title":"Participant observation: A guide for fieldworkers","author":"K. DeWalt","year":"2010","unstructured":"DeWalt, K. M, & DeWalt, B. R. (2010). Participant observation: A guide for fieldworkers. Maryland, US: Rowman & Littlefield."},{"key":"221_CR18","doi-asserted-by":"publisher","first-page":"197","DOI":"10.1093\/oxfordhb\/9780190067397.013.11","volume-title":"The Oxford handbook of ethics of AI","author":"N. Diakopoulos","year":"2020","unstructured":"Diakopoulos, N. (2020). Accountability, transparency, and algorithms. In M. D. Dubber, F. Pasquale, & S. Das (Eds.), The Oxford handbook of ethics of AI (pp. 197\u2013213). Oxford University Press."},{"issue":"2","key":"221_CR19","doi-asserted-by":"publisher","first-page":"89","DOI":"10.1109\/TTS.2020.2992344","volume":"1","author":"P. Drozdowski","year":"2020","unstructured":"Drozdowski, P., Rathgeb, C., Dantcheva, A., Damer, N., & Busch, C. (2020). Demographic bias in biometrics: A survey on an emerging challenge. IEEE Transactions on Technology and Society, 1(2), 89\u2013103.","journal-title":"IEEE Transactions on Technology and Society"},{"key":"221_CR20","unstructured":"European Commission. (2020). AI high-level expert group - assessment list for trustworthy artificial intelligence. https:\/\/futurium.ec.europa.eu\/en\/european-ai-alliance\/document\/ai-hleg-assessment-list-trustworthy-artificial-intelligence-altai?language=fr."},{"key":"221_CR21","unstructured":"European Law Institute. (2022). Model rules on impact assessment of algorithmic decision-making systems used by public administration. https:\/\/www.europeanlawinstitute.eu\/fileadmin\/user_upload\/p_eli\/Publications\/ELI_Model_Rules_on_Impact_Assessment_of_ADMSs_Used_by_Public_Administration.pdf."},{"issue":"4\u20135","key":"221_CR22","doi-asserted-by":"publisher","first-page":"663","DOI":"10.1080\/09614520701469955","volume":"17","author":"J. Fox","year":"2007","unstructured":"Fox, J. (2007). The uncertain relationship between transparency and accountability. Development in Practice, 17(4\u20135), 663\u2013671. https:\/\/doi.org\/10.1080\/09614520701469955.","journal-title":"Development in Practice"},{"key":"221_CR23","unstructured":"Gertler, N. (2023). Hacking AI governance: Exploring the democratic potential of Canada\u2019s algorithmic impact assessment [Masters thesis]. Concordia University. https:\/\/spectrum.library.concordia.ca\/id\/eprint\/992742\/. Montreal, Canada."},{"key":"221_CR24","unstructured":"Gertler, N. (2024). Canada\u2019s algorithmic impact assessment. In F. McKelvey, S. Toupin, & J. Roberge (Eds.), Northern lights and silicon dreams: AI governance in Canada (2011-2022) (pp. 31\u201341). Montreal, Canada. https:\/\/www.amo-oma.ca\/en\/ai-policy-report\/."},{"key":"221_CR25","unstructured":"Gonzalez, V., & Salomon, G. (2025). Orders to leave the country - some for US citizens - sow confusion among immigrants. Associated Press. https:\/\/apnews.com\/article\/trump-immigration-cbp-one-asylum-4a3aae0453d17b6dfc2c9b590f19497e."},{"key":"221_CR26","volume-title":"Algorithmic impact assessment","author":"Government of Canada","year":"2022","unstructured":"Government of Canada. (2022). Algorithmic Impact Assessment. Version 0.9.1. Treasury Board of Canada Secretariat. https:\/\/github.com\/canada-ca\/aia-eia-js\/milestones."},{"key":"221_CR27","unstructured":"Government of Canada. (2023). Directive on automated decision-making. Treasury board of Canada secretariat. https:\/\/www.tbs-sct.canada.ca\/pol\/doc-eng.aspx?id=32592#appC."},{"key":"221_CR28","unstructured":"Government of Canada. (2024a). Algorithmic impact assessment tool. Treasury board of Canada secretariat. https:\/\/www.canada.ca\/en\/government\/system\/digital-government\/digital-government-innovations\/responsible-use-ai\/algorithmic-impact-assessment.html."},{"key":"221_CR29","unstructured":"Government of Canada. (2024b). Algorithmic impact assessment version 0.10.0. Treasury board of Canada secretariat. https:\/\/canada-ca.github.io\/aia-eia-js\/."},{"key":"221_CR30","unstructured":"Government of Canada. (2024c). Guide on the scope of the directive on automated decision-making. Treasury board of Canada secretariat. https:\/\/www.canada.ca\/en\/government\/system\/digital-government\/digital-government-innovations\/responsible-use-ai\/guide-scope-directive-automated-decision-making.html."},{"key":"221_CR31","volume-title":"Open government portal: \u201calgorithmic impact assessments\u201d","author":"Government of Canada","year":"2025","unstructured":"Government of Canada. (2025). Open Government Portal: \u201cAlgorithmic Impact Assessments\u201d. Treasury Board of Canada Secretariat. https:\/\/search.open.canada.ca\/opendata\/?sort=metadata_created+desc%26search_text=%22algorithmic+impact+assessment%22%26;page=1."},{"issue":"2","key":"221_CR32","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1007\/s44206-022-00017-z","volume":"1","author":"A. Hasan","year":"2022","unstructured":"Hasan, A., Brown, S., Davidovic, J., Lange, B., & Regan, M. (2022). Algorithmic bias and risk assessments: Lessons from practice. Digital Society, 1(2), 14. https:\/\/doi.org\/10.1007\/s44206-022-00017-z.","journal-title":"Digital Society"},{"key":"221_CR33","unstructured":"Heisler, N. (2022). Standards for the control of algorithmic bias in the Canadian administrative context (Master\u2019s thesis, University of Waterloo). https:\/\/uwspace.uwaterloo.ca\/items\/4939baca-d81c-4531-934d-83e22c83cd49. Waterloo, Canada."},{"key":"221_CR34","doi-asserted-by":"crossref","unstructured":"Hsieh, H. F., & Shannon, S. E. (2005). Three approaches to qualitative content analysis. The Qualitative Health Research, 15(9), 1277\u20131288.","DOI":"10.1177\/1049732305276687"},{"key":"221_CR35","unstructured":"Institute for the Future of Work. (2022). International learnings on algorithmic impact assessments (panel). https:\/\/www.youtube.com\/watch?v=kFti_oKZFsw."},{"key":"221_CR36","unstructured":"Institute for the Future of Work. (2023). Good Work AIA. https:\/\/www.ifow.org\/publications\/good-work-algorithmic-impact-assessment-an-approach-for-worker-involvement."},{"key":"221_CR37","unstructured":"Karadeglija, A. (2024). Federal government use of AI in hundreds of initiatives revealed by new research database. CBC news. https:\/\/www.cbc.ca\/news\/politics\/federal-government-used-ai-1.7170307."},{"key":"221_CR38","unstructured":"Karlin, M. (2018). Deploying AI responsibly in government. Policy options. https:\/\/policyoptions.irpp.org\/magazines\/february-2018\/deploying-ai-responsibly-in-government\/."},{"key":"221_CR39","unstructured":"Kaye, K. (2024). How Canada\u2019s algorithmic impact assessment process and algorithm has evolved. World privacy forum. Retrieved from https:\/\/www.worldprivacyforum.org\/wp-content\/uploads\/2024\/08\/WPF_AI_Governance_Canada_AIA_August2024_fs.pdf."},{"issue":"6","key":"221_CR40","doi-asserted-by":"publisher","first-page":"101976","DOI":"10.1016\/j.telpol.2020.101976","volume":"44","author":"M. Kuziemski","year":"2020","unstructured":"Kuziemski, M., & Misuraca, G. (2020). AI governance in the public sector: Three tales from the frontiers of automated decision-making in democratic settings. Telecommunications Policy, 44(6), 101976. https:\/\/doi.org\/10.1016\/j.telpol.2020.101976.","journal-title":"Telecommunications Policy"},{"issue":"182","key":"221_CR41","doi-asserted-by":"publisher","first-page":"26 pp","DOI":"10.1145\/3359284","volume":"3","author":"M. K. Lee","year":"2019","unstructured":"Lee, M. K., Jain, A., Cha, H. J., Ojha, S., & Kusbit, K. (2019, November). Procedural justice in algorithmic fairness: Leveraging transparency and outcome control for fair algorithmic mediation. Proceedings of the ACM Human-Computer Interaction, 3(182), 26 pp. https:\/\/doi.org\/10.1145\/3359284.","journal-title":"Proceedings of the ACM Human-Computer Interaction"},{"key":"221_CR42","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1007\/978-1-4419-6536-3_7","volume-title":"Comparative E-Government","author":"L. F. Luna-Reyes","year":"2010","unstructured":"Luna-Reyes, L. F., Pardo, T. A., Gil-Garcia, J. R., Navarrete, C., Zhang, J., & Mellouli, S. (2010). Digital government in North America: A comparative analysis of policy and program priorities in Canada, Mexico, and the United States. In C. Reddick (Ed.), Comparative E-Government (pp. 139\u2013160). Integrated Series in Information Systems, 25. New York, NY: Springer. https:\/\/doi.org\/10.1007\/978-1-4419-6536-3_7."},{"key":"221_CR43","doi-asserted-by":"crossref","unstructured":"Metcalf, J., Moss, E., Watkins, E. A., Singh, R., & Elish, M. C. (2021, March). Algorithmic impact assessments and accountability: The co-construction of impacts. In Proceedings of the 2021 ACM Conference On Fairness, Accountability, and Transparency (pp. 735\u2013746). https:\/\/dl.acm.org\/doi\/abs\/10.1145\/3442188.3445935.","DOI":"10.1145\/3442188.3445935"},{"issue":"2128","key":"221_CR44","doi-asserted-by":"publisher","first-page":"20170357","DOI":"10.1098\/rsta.2017.0357","volume":"376","author":"S. J. Mikhaylov","year":"2018","unstructured":"Mikhaylov, S. J., Esteve, M., & Campion, A. (2018). Artificial intelligence for the public sector: Opportunities and challenges of cross-sector collaboration. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 376(2128), 20170357. https:\/\/doi.org\/10.1098\/rsta.2017.0357.","journal-title":"Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences"},{"key":"221_CR45","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1007\/s11023-021-09557-8","volume":"31","author":"J. M\u00f6kander","year":"2021","unstructured":"M\u00f6kander, J., & Floridi, L. (2021). Ethics-based auditing to develop trustworthy AI. Minds and Machines, 31, 323\u2013327.","journal-title":"Minds and Machines"},{"key":"221_CR46","doi-asserted-by":"crossref","unstructured":"Moss, E., Watkins, E. A., Singh, R., Elish, M. C., & Metcalf, J. (2021). Assembling accountability: Algorithmic impact assessment for the public interest. Data & society research institute. https:\/\/datasociety.net\/library\/assembling-accountability-algorithmic-impact-assessment-for-the-public-interest\/.","DOI":"10.2139\/ssrn.3877437"},{"key":"221_CR47","first-page":"773","volume":"34","author":"D. K. Mulligan","year":"2019","unstructured":"Mulligan, D. K., & Bamberger, K. A. (2019). Procurement as policy: Administrative process for machine learning. Berkeley Technology Law Journal, 34, 773.","journal-title":"Berkeley Technology Law Journal"},{"key":"221_CR48","volume-title":"Artificial intelligence risk management framework (AI RMF 1.0)","author":"National Institute of Standards and Technology","year":"2023","unstructured":"National Institute of Standards and Technology. (2023). Artificial intelligence risk management framework (AI RMF 1.0). U.S. Department of Commerce. https:\/\/nvlpubs.nist.gov\/nistpubs\/ai\/NIST.AI.100-1.pdf."},{"key":"221_CR49","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1093\/oxfordhb\/9780199641253.013.0002","volume-title":"The Oxford handbook for public accountability","author":"S. V. Patil","year":"2014","unstructured":"Patil, S. V., Vieider, F., & Tetlock, P. E. (2014). Process versus outcome accountability. In M. A. P. Bovens, R. E. Goodin, & T. Schillemans (Eds.), The Oxford handbook for public accountability (pp. 69\u201389). https:\/\/doi.org\/10.1093\/oxfordhb\/9780199641253.013.0002."},{"key":"221_CR50","unstructured":"Public Policy Forum. (1998). Innovation in the federal government: The risk not taken. background document for a roundtable discussion to be held on behalf of the office of the Auditor General. https:\/\/www.oag-bvg.gc.ca\/internet\/english\/meth_gde_e_10193.html."},{"key":"221_CR51","unstructured":"Quenneville, G. (2022). Meet the librarian in rural Vermont helping stranded travellers with the ArriveCAN app. CBC news. https:\/\/www.cbc.ca\/news\/politics\/vermont-librarian-arrivecan-app-1.6562293\/."},{"key":"221_CR52","unstructured":"Reeveley, D. (2021). Federal rules on AI too narrow and risk \u2018damaging public trust\u2019: Internal review. The logic. https:\/\/thelogic.co\/news\/federal-rules-on-ai-too-narrow-and-risk-damaging-public-trust-internal-review."},{"key":"221_CR53","unstructured":"Reisman, D., Schultz, J., Crawford, K., & Whittaker, M. (2018). Algorithmic impact assessments: A practical framework for public agency accountability. AI Now Institute. https:\/\/ainowinstitute.org\/publication\/algorithmic-impact-assessments-report-2."},{"key":"221_CR54","volume-title":"Artificial intelligence and equality initiative","author":"N. Sambuli","year":"2021","unstructured":"Sambuli, N. (2021). Five challenges with multistakeholder initiatives on AI. In Artificial intelligence and equality initiative. Carnegie Council for Ethics and International Affairs. https:\/\/www.carnegiecouncil.org\/media\/article\/five-challenges-with-multistakeholder-initiatives-on-ai."},{"key":"221_CR55","unstructured":"Sandvig, C., Hamilton, K., Karahalios, K., & Langbort, C. (2014). Auditing algorithms: Research methods for detecting discrimination on internet platforms. In Data and Discrimination: Converting Critical Concerns into Productive Inquiry. Preconference at the 64th Annual Meeting of the International Communication Association, Seattle, WA, USA. https:\/\/websites.umich.edu\/%7Ecsandvig\/research\/Auditing%20Algorithms%20-%20Sandvig%20-%20ICA%202014%20Data%20and%20Discrimination%20Preconference.pdf."},{"key":"221_CR56","unstructured":"Schneider, K. (2024). RCMP\u2019s proposed AI surveillance system for holding cells called \u2018dehumanizing\u2019 and \u2018intrusive\u2019 by experts. Investigative journalism foundation. https:\/\/theijf.org\/rcmp-ai-surveillance-system."},{"issue":"1","key":"221_CR57","first-page":"117","volume":"35","author":"A. D. Selbst","year":"2021","unstructured":"Selbst, A. D. (2021). An institutional view of algorithmic impact assessments. Harvard Journal of Law & Technology (Harvard JOLT), 35(1), 117\u2013192. https:\/\/jolt.law.harvard.edu\/assets\/articlePDFs\/v35\/Selbst-An-Institutional-View-of-Algorithmic-Impact-Assessments.pdf.","journal-title":"Harvard Journal of Law & Technology (Harvard JOLT)"},{"key":"221_CR58","doi-asserted-by":"publisher","unstructured":"Sieber, R. E. (2022). Raw political power in civic engagement with AI. In A. Brandusescu & J. Reia (Eds.), Artificial intelligence in the city: Building civic engagement and public trust (pp. 17\u201318). https:\/\/doi.org\/10.18130\/9kar-xn17.","DOI":"10.18130\/9kar-xn17"},{"key":"221_CR59","doi-asserted-by":"publisher","unstructured":"Sieber, R. E., Brandusescu, A., Adu-Daako, A., & Sangiambut, S. (2024). Who are the publics engaging in AI? Public Understanding of Science, 33(5), 634\u2013653. https:\/\/doi.org\/10.1177\/09636625231219853.","DOI":"10.1177\/09636625231219853"},{"key":"221_CR60","doi-asserted-by":"publisher","unstructured":"Sieber, R., Brandusescu, A., Sangiambut, S., & Adu-Daako, A. (2025). What is civic participation in artificial intelligence? Environment and Planning B: Urban Analytics and City Science, 52(6), 1388\u20131406. https:\/\/doi.org\/10.1177\/23998083241296200.","DOI":"10.1177\/23998083241296200"},{"key":"221_CR61","doi-asserted-by":"crossref","unstructured":"Sloane, M., & Moss, E. (2023). Assessing the assessment: Comparing algorithmic impact assessments and AI audits. SSRN, 14. https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=4486259.","DOI":"10.2139\/ssrn.4486259"},{"key":"221_CR62","doi-asserted-by":"publisher","first-page":"253","DOI":"10.1016\/0016-7185(84)90036-8","volume":"2","author":"L. G. Smith","year":"1984","unstructured":"Smith, L. G. (1984). Public participation in policy making: The state-of-the-art in Canada. Geoforum, 2, 253\u2013259.","journal-title":"Geoforum"},{"key":"221_CR63","doi-asserted-by":"publisher","first-page":"12799","DOI":"10.1007\/s10462-023-10420-8","volume":"56","author":"B. C. Stahl","year":"2023","unstructured":"Stahl, B. C., Antoniou, J., Bhalla, N. B., Jansen, L., Lindqvist, P., Kirichenko, B., Marchal, A., Rodrigues, S., Santiago, R., Warso, N., Z, & Wright, D. (2023). A systematic review of artificial intelligence impact assessments. Artificial Intelligence Review, 56, 12799\u201312831. https:\/\/doi.org\/10.1007\/s10462-023-10420-8.","journal-title":"Artificial Intelligence Review"},{"issue":"1","key":"221_CR64","doi-asserted-by":"publisher","first-page":"16 pp","DOI":"10.1177\/20539517221100449","volume":"9","author":"A. Waldman","year":"2022","unstructured":"Waldman, A., & Martin, K. (2022). Governing algorithmic decisions: The role of decision importance and governance on perceived legitimacy of algorithmic decisions. Big Data & Society, 9(1), 16 pp. https:\/\/doi.org\/10.1177\/20539517221100449.","journal-title":"Big Data & Society"},{"key":"221_CR65","unstructured":"Waldman, A. E. (2019). Power, process, and automated decision-making. Fordham Law Review, 88, 613. https:\/\/ir.lawnet.fordham.edu\/flr\/vol88\/iss2\/9\/."},{"key":"221_CR66","doi-asserted-by":"publisher","unstructured":"Watkins, E. A., Moss, E., Metcalf, J., Singh, R., & Elish, M. C. (2021). Governing algorithmic systems with impact assessments: Six observations. In Proceedings of the 2021 AAAI\/ACM Conference on AI, Ethics, and Society (pp. 1010\u20131022). New York City, NY, USA. https:\/\/doi.org\/10.1145\/3461702.3462580.","DOI":"10.1145\/3461702.3462580"},{"issue":"29","key":"221_CR67","first-page":"36 pp","volume":"3","author":"A. Wernick","year":"2024","unstructured":"Wernick, A. (2024). Impact assessment as a legal design pattern-A \u201ctimeless way\u201d of managing future risks? Digital Society, 3(29), 36 pp.","journal-title":"Digital Society"},{"key":"221_CR68","doi-asserted-by":"crossref","unstructured":"Wieringa, M. (2020, January). What to account for when accounting for algorithms: A systematic literature review on algorithmic accountability. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (pp. 1\u201318).","DOI":"10.1145\/3351095.3372833"}],"container-title":["Digital Society"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44206-025-00221-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44206-025-00221-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44206-025-00221-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,3]],"date-time":"2025-09-03T15:27:12Z","timestamp":1756913232000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44206-025-00221-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8]]},"references-count":68,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["221"],"URL":"https:\/\/doi.org\/10.1007\/s44206-025-00221-7","relation":{},"ISSN":["2731-4650","2731-4669"],"issn-type":[{"value":"2731-4650","type":"print"},{"value":"2731-4669","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,8]]},"assertion":[{"value":"17 October 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 July 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 August 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"There is no conflict of interest. While we have been asked to consult with TBS (the developers of the AIA), we have always done this at arms-length and TBS comes to us because we are critics. We are not paid as consultants nor employed by the government.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"64"}}