{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T20:51:02Z","timestamp":1776113462346,"version":"3.50.1"},"reference-count":142,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2022,10,2]],"date-time":"2022-10-02T00:00:00Z","timestamp":1664668800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,10,2]],"date-time":"2022-10-02T00:00:00Z","timestamp":1664668800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["AI &amp; Soc"],"published-print":{"date-parts":[[2024,6]]},"DOI":"10.1007\/s00146-022-01572-2","type":"journal-article","created":{"date-parts":[[2022,10,2]],"date-time":"2022-10-02T05:02:20Z","timestamp":1664686940000},"page":"1213-1227","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":46,"title":["Public procurement of artificial intelligence systems: new risks and future proofing"],"prefix":"10.1007","volume":"39","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3068-3108","authenticated-orcid":false,"given":"Merve","family":"Hickok","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,2]]},"reference":[{"key":"1572_CR1","unstructured":"A Civil Society Statement (2021). An EU artificial intelligence act for fundamental rights. https:\/\/edri.org\/wp-content\/uploads\/2021\/12\/Political-statement-on-AI-Act.pdf"},{"key":"1572_CR2","unstructured":"ACLU (2022a) Three key problems with the government\u2019s use of a flawed facial recognition service. ACLU Florida. https:\/\/www.aclufl.org\/en\/news\/three-key-problems-governments-use-flawed-facial-recognition-service"},{"key":"1572_CR3","unstructured":"ACLU (2022b) Settlement secures the right to consideration of release for people arrested by ICE in New York. ACLU New York. https:\/\/www.nyclu.org\/en\/press-releases\/settlement-secures-right-consideration-release-people-arrested-ice-new-york"},{"key":"1572_CR4","unstructured":"Ada Lovelace Institute, AI Now Institute and Open Government Partnership (2021) Algorithmic accountability for the public sector. https:\/\/www.opengovpartnership.org\/documents\/algorithmic-accountability-public-sector\/"},{"key":"1572_CR5","doi-asserted-by":"publisher","first-page":"52138","DOI":"10.1109\/ACCESS.2018.2870052","volume":"6","author":"A Adadi","year":"2018","unstructured":"Adadi A, Berrada M (2018) Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE Access 6:52138\u201352160. https:\/\/doi.org\/10.1109\/ACCESS.2018.2870052","journal-title":"IEEE Access"},{"key":"1572_CR6","unstructured":"AI Now Institute NYU (2018) Automated decision systems: examples of government use cases. https:\/\/ainowinstitute.org\/nycadschart.pdf"},{"key":"1572_CR7","unstructured":"Aleazis H, Haskins C (2020) DHS authorities are buying moment-by-moment geolocation cellphone data to track people. BuzzFeed News. https:\/\/www.buzzfeednews.com\/article\/hamedaleaziz\/ice-dhs-cell-phone-data-tracking-geolocation"},{"key":"1572_CR8","unstructured":"AlgorithmWatch (2020) Automating society report 2020. https:\/\/algorithmwatch.org\/en\/automating-society-2020\/"},{"issue":"3","key":"1572_CR9","doi-asserted-by":"publisher","first-page":"973","DOI":"10.1177\/1461444816676645","volume":"20","author":"M Ananny","year":"2018","unstructured":"Ananny M, Crawford K (2018) Seeing without knowing. New Media Soc 20(3):973\u2013989. https:\/\/doi.org\/10.1177\/1461444816676645","journal-title":"New Media Soc"},{"key":"1572_CR10","unstructured":"Angwin J, Larson J, Mattu S, Kirchner L (2016) Machine bias: there's software used across the country to predict future criminals. And it's biased against blacks. ProPublica. https:\/\/www.propublica.org\/article\/machine\u2010bias\u2010risk\u2010assessments\u2010in\u2010criminal\u2010sentencing"},{"key":"1572_CR11","unstructured":"Arkansas Department of Human Services v. Ledgerwood, 530. Arkansas Supreme Court (2017) Casetext. https:\/\/casetext.com\/case\/ark-dept-of-human-servs-v-ledgerwood-1"},{"key":"1572_CR12","doi-asserted-by":"crossref","unstructured":"Barocas S, Selbst AD (2016) Big data's disparate impact. California Law Review 104: 671. https:\/\/www.californialawreview.org\/wp-content\/uploads\/2016\/06\/2Barocas-Selbst.pdf","DOI":"10.2139\/ssrn.2477899"},{"key":"1572_CR13","unstructured":"Barry v. Lyon, 834. 6th Circuit (2016) Casetext. https:\/\/casetext.com\/case\/barry-v-lyon-2"},{"key":"1572_CR14","unstructured":"BBC (2020) A-levels and GCSEs: How did the exam algorithm work? https:\/\/www.bbc.com\/news\/explainers-53807730"},{"key":"1572_CR15","doi-asserted-by":"crossref","unstructured":"Bender E, Friedman B (2018) Data statements for natural language processing: toward mitigating system bias and enabling better science. Trans Assoc Comput Linguist [Online], 6: 587\u2013604. https:\/\/aclanthology.org\/Q18-1041\/","DOI":"10.1162\/tacl_a_00041"},{"key":"1572_CR16","doi-asserted-by":"publisher","DOI":"10.1177\/0049124118782533","author":"RA Berk","year":"2018","unstructured":"Berk RA, Heidari H, Jabbari S, Kearns M, Roth A (2018) Fairness in criminal justice risk assessments: the state of the art. Sociol Methods Res. https:\/\/doi.org\/10.1177\/0049124118782533","journal-title":"Sociol Methods Res"},{"key":"1572_CR17","unstructured":"Biddle S (2021) LexisNexis to provide giant database of personal information to ICE. Intercept. https:\/\/theintercept.com\/2021\/04\/02\/ice-database-surveillance-lexisnexis"},{"key":"1572_CR18","doi-asserted-by":"publisher","unstructured":"Birhane A, Kalluri P, Card D, Agnew W, Dotan R, Bao M (2022) The values encoded in machine learning research. FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency. June 2022.173\u2013184. https:\/\/doi.org\/10.1145\/3531146.3533083","DOI":"10.1145\/3531146.3533083"},{"key":"1572_CR19","unstructured":"Black C (2021) Revealed: data giant given \u2018emergency\u2019 covid contract had been wooing NHS for months. The Bureau of Investigative Journalism. https:\/\/www.thebureauinvestigates.com\/stories\/2021-02-24\/revealed-data-giant-given-emergency-covid-contract-had-been-wooing-nhs-for-months"},{"issue":"4","key":"1572_CR20","doi-asserted-by":"publisher","first-page":"447","DOI":"10.1111\/j.1468-0386.2007.00378.x","volume":"13","author":"M Bovens","year":"2007","unstructured":"Bovens M (2007) Analysing and assessing accountability: a conceptual framework. Eur Law J 13(4):447\u2013468. https:\/\/doi.org\/10.1111\/j.1468-0386.2007.00378.x","journal-title":"Eur Law J"},{"key":"1572_CR21","doi-asserted-by":"publisher","DOI":"10.1093\/oso\/9780190684099.001.0001","volume-title":"Predict and surveil: data, discretion, and the future of policing","author":"S Brayne","year":"2020","unstructured":"Brayne S (2020) Predict and surveil: data, discretion, and the future of policing, 1st edn. Oxford University Press","edition":"1"},{"key":"1572_CR22","unstructured":"Brewster T (2021) These companies track millions of cars\u2014immigration and border police have been grabbing their data. Forbes. https:\/\/www.forbes.com\/sites\/thomasbrewster\/2021\/04\/01\/these-companies-track-millions-of-cars-immigration-and-border-police-have-been-grabbing-their-data\/"},{"key":"1572_CR23","doi-asserted-by":"publisher","unstructured":"Brown S, Carrier R, Hickok M, Smith AL (2021). Bias mitigation in data sets. https:\/\/doi.org\/10.31235\/osf.io\/z8qrb","DOI":"10.31235\/osf.io\/z8qrb"},{"key":"1572_CR24","unstructured":"Buolamwini J (2022) The IRS should stop using facial recognition. The Atlantic. https:\/\/www.theatlantic.com\/ideas\/archive\/2022\/01\/irs-should-stop-using-facial-recognition\/621386\/"},{"key":"1572_CR25","first-page":"1","volume":"81","author":"J Buolamwini","year":"2018","unstructured":"Buolamwini J, Gebru T (2018) Gender shades: intersectional accuracy disparities in commercial gender classification. Proc Mach Learn Res 81:1\u201315","journal-title":"Proc Mach Learn Res"},{"issue":"5","key":"1572_CR26","doi-asserted-by":"publisher","first-page":"825","DOI":"10.1111\/puar.13293","volume":"81","author":"M Busuioc","year":"2021","unstructured":"Busuioc M (2021) Accountable artificial intelligence: holding algorithms to account. Public Adm Rev 81(5):825\u2013836. https:\/\/doi.org\/10.1111\/puar.13293","journal-title":"Public Adm Rev"},{"key":"1572_CR27","unstructured":"Cahoo v. SAS Analytics Inc., 912. 6th Circuit (2019). Casetext. https:\/\/casetext.com\/case\/cahoo-v-sas-analytics-inc"},{"issue":"4","key":"1572_CR28","first-page":"797","volume":"70","author":"R Calo","year":"2021","unstructured":"Calo R, Citron DK (2021) The automated administrative state: a crisis of legitimacy. Emory Law J 70(4):797\u2013845","journal-title":"Emory Law J"},{"key":"1572_CR29","unstructured":"Campbell AF (2018) How tech employees are pushing Silicon Valley to put ethics before profit. Vox. https:\/\/www.vox.com\/technology\/2018\/10\/18\/17989482\/google-amazon-employee-ethics-contracts"},{"key":"1572_CR30","unstructured":"Center for AI and Digital Policy (2021) AI & Democratic Values Index 2020. www.caidp.org\/reports\/aidv-2020\/"},{"key":"1572_CR31","unstructured":"Center for AI and Digital Policy (2022) AI & Democratic Values Index 2021. www.caidp.org\/reports\/aidv-2021\/"},{"key":"1572_CR32","unstructured":"Central Digital and Data Office (2021) Algorithmic transparency standard. https:\/\/www.gov.uk\/government\/collections\/algorithmic-transparency-standard"},{"issue":"2017","key":"1572_CR33","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1089\/big.2016.0047","volume":"2","author":"A Chouldechova","year":"2017","unstructured":"Chouldechova A (2017) Fair prediction with disparate impact: a study of bias in recidivism prediction instruments. Big Data 2(2017):153\u2013163. https:\/\/doi.org\/10.1089\/big.2016.0047","journal-title":"Big Data"},{"key":"1572_CR34","unstructured":"City of Amsterdam (2020) Amsterdam algorithm register. https:\/\/algoritmeregister.amsterdam.nl\/en\/ai-register"},{"key":"1572_CR35","unstructured":"City of New York (2020) Public oversight of surveillance technologies act. https:\/\/legistar.council.nyc.gov\/LegislationDetail.aspx?ID=3343878&GUID=996ABB2A-9F4C-4A32-B081-D6F24AB954A0"},{"key":"1572_CR36","unstructured":"Coglianese C, Lampmann E (2021) Contracting for algorithmic accountability, Adm Law Rev Accord, vol. 6, 175. https:\/\/scholarship.law.upenn.edu\/faculty_scholarship\/2311\/"},{"key":"1572_CR37","unstructured":"Coglianese C, Lehr D (2017) Regulating by robot: administrative decision making in the machine-learning era. Faculty Scholarship at Penn Carey Law. 1734. https:\/\/scholarship.law.upenn.edu\/faculty_scholarship\/1734"},{"key":"1572_CR38","doi-asserted-by":"publisher","unstructured":"Cooper AF, Laufer B, Moss E, Nissenbaum H (2022) Accountability in an algorithmic society: relationality, responsibility, and robustness in machine learning. FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency. June 2022. 864\u2013876. https:\/\/doi.org\/10.1145\/3531146.3533150","DOI":"10.1145\/3531146.3533150"},{"key":"1572_CR39","unstructured":"Council of Europe (2021) Possible elements of a legal framework on artificial intelligence, based on the Council of Europe\u2019s standards on human rights, democracy and the rule of law. https:\/\/www.coe.int\/en\/web\/artificial-intelligence\/work-in-progress"},{"key":"1572_CR40","unstructured":"Crump C (2016) Surveillance policy making by procurement, 91 Washington Law Review. 1595 (2016). https:\/\/digitalcommons.law.uw.edu\/wlr\/vol91\/iss4\/17"},{"key":"1572_CR41","unstructured":"Department of Justice (2004) FOIA guide, 2004 edition. https:\/\/www.justice.gov\/archives\/oip\/foia-guide-2004-edition-exemption-4"},{"key":"1572_CR42","volume-title":"Algorithmic accountability reporting: on the investigation of black boxes","author":"N Diakopoulos","year":"2014","unstructured":"Diakopoulos N (2014) Algorithmic accountability reporting: on the investigation of black boxes. Tow Center for Digital Journalism Publications, Columbia Journalism School"},{"key":"1572_CR43","unstructured":"District Court of Hague (2020). SyRI. https:\/\/perma.cc\/DS89-K477. English explanation: library of Congress (March 13, 2020). Court prohibits government's use of AI software to detect welfare fraud. https:\/\/www.loc.gov\/item\/global-legal-monitor\/2020-03-13\/netherlands-court-prohibits-governments-use-of-ai-software-to-detect-welfare-fraud\/"},{"key":"1572_CR44","doi-asserted-by":"publisher","unstructured":"Dobbe R, Dean S, Gilbert TK, Kohli N (2018) A broader view on bias in automated decision-making: reflecting on epistemology and dynamics. Workshop on Fairness, Accountability and Transparency in Machine Learning during ICML 2018, Stockholm, Sweden. https:\/\/doi.org\/10.48550\/arXiv.1807.00553","DOI":"10.48550\/arXiv.1807.00553"},{"key":"1572_CR45","volume-title":"Digital era governance: IT corporations, the state, and e-government","author":"P Dunleavy","year":"2007","unstructured":"Dunleavy P, Margetts HZ, Bastow S, Tinkler J (2007) Digital era governance: IT corporations, the state, and e-government. Oxford University Press"},{"key":"1572_CR46","unstructured":"Dutch Parliament (2020) Eindverslag onderzoek kinderopvangtoeslag overhandigd. https:\/\/www.tweedekamer.nl\/nieuws\/kamernieuws\/eindverslag-onderzoek-kinderopvangtoeslag-overhandigd?msclkid=f1d677c5ae8311ecaaa202cbd2ef5f6d"},{"key":"1572_CR47","unstructured":"Dutch Parliament (2022) otie van de leden Bouchallikh en Dekker-Abdulaziz over verplichte impact assessments voorafgaand aan het inzetten van algoritmen voor evaluaties van of beslissingen over mensen. https:\/\/www.tweedekamer.nl\/kamerstukken\/moties\/detail?id=2022Z06024&did=2022D12329"},{"key":"1572_CR48","doi-asserted-by":"publisher","unstructured":"Dwork C, Hardt M, Pitassi T, Reingold O, Zemel RS (2012) Fairness through awareness. ITCS '12: Proc. of the 3rd Innovations in Theoretical Computer Science Conference, 214\u2013226. https:\/\/doi.org\/10.1145\/2090236.2090255","DOI":"10.1145\/2090236.2090255"},{"issue":"2","key":"1572_CR49","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1177\/0093854818811379","volume":"46","author":"L Eckhouse","year":"2019","unstructured":"Eckhouse L, Lum K, Conti-Cook C, Ciccolini J (2019) Layers of bias: a unified approach for understanding problems with risk assessment. Crim Justice Behav 46(2):185\u2013209. https:\/\/doi.org\/10.1177\/0093854818811379","journal-title":"Crim Justice Behav"},{"key":"1572_CR50","doi-asserted-by":"crossref","unstructured":"Engstrom DF, Ho DE, Sharkey CM, Cu\u00e9llar M (2020) Government by algorithm: artificial intelligence in federal administrative agencies. Administrative Conference of the United States. https:\/\/www-cdn.law.stanford.edu\/wp-content\/uploads\/2020\/02\/ACUS-AI-Report.pdf","DOI":"10.2139\/ssrn.3551505"},{"key":"1572_CR51","unstructured":"Ensign DL, Friedler SA, Neville S, Scheidegger CE, Venkatasubramanian S (2018) Runaway feedback loops in predictive policing. FAT."},{"key":"1572_CR52","volume-title":"Automating inequality: how high-tech tools profile, police, and punish the poor","author":"V Eubanks","year":"2018","unstructured":"Eubanks V (2018) Automating inequality: how high-tech tools profile, police, and punish the poor. St. Martin\u2019s Press"},{"key":"1572_CR53","unstructured":"European Parliament (2020) Parliamentary question: E-000173\/2020h https:\/\/www.europarl.europa.eu\/doceo\/document\/E-9-2020h-000173-ASW_EN.html"},{"key":"1572_CR54","unstructured":"European Commission (2021) Proposal for a regulation of the European parliament and of the council laying down harmonised rules on artificial intelligence (artificial intelligence act) and amending certain union legislative acts (2021) https:\/\/eur-lex.europa.eu\/legal-content\/EN\/TXT\/?uri=CELEX%3A52021PC0206"},{"key":"1572_CR55","unstructured":"Executive Order of President 13859 (February 11, 2019). Maintaining American Leadership in Artificial Intelligence. https:\/\/www.federalregister.gov\/documents\/2019\/02\/14\/2019-02544\/maintaining-american-leadership-in-artificial-intelligence"},{"key":"1572_CR56","unstructured":"Executive Order of President 13960 (2020). Promoting the use of trustworthy artificial intelligence in the federal government. https:\/\/www.federalregister.gov\/documents\/2020\/12\/08\/2020-27065\/promoting-the-use-of-trustworthy-artificial-intelligence-in-the-federal-government"},{"key":"1572_CR57","unstructured":"Faife C (2021). Utility companies will no longer share data with ICE \u2014 but many loopholes remain. The Verge. https:\/\/www.theverge.com\/2021\/12\/9\/22826271\/utilities-ice-data-sharing-thomson-wyden"},{"key":"1572_CR58","unstructured":"Fields-White M, Graubard V, Rodr\u00edguez \u00c1lvarez A, Zeichner N, Robertson C (2020) Unpacking inequities in unemployment insurance. New America. https:\/\/www.newamerica.org\/pit\/reports\/unpacking-inequities-unemployment-insurance\/a-focus-on-fraud-over-accessibility-the-punitive-design-of-ui"},{"issue":"4","key":"1572_CR59","doi-asserted-by":"publisher","first-page":"403","DOI":"10.1016\/0957-4174(94)E0032-P","volume":"8","author":"DE Forsythe","year":"1995","unstructured":"Forsythe DE (1995) Using ethnography in the design of an explanation system. Expert Syst Appl 8(4):403\u2013417. https:\/\/doi.org\/10.1016\/0957-4174(94)E0032-P","journal-title":"Expert Syst Appl"},{"issue":"4","key":"1572_CR60","doi-asserted-by":"publisher","first-page":"136","DOI":"10.1145\/3433949","volume":"64","author":"SA Friedler","year":"2021","unstructured":"Friedler SA, Scheidegger C, Venkatasubramanian S (2021) The (Im)possibility of fairness: different value systems require different mechanisms for fair decision making. Commun ACM 64(4):136. https:\/\/doi.org\/10.1145\/3433949","journal-title":"Commun ACM"},{"key":"1572_CR61","unstructured":"De La Garza A (2020) States' automated systems are trapping citizens in bureaucratic nightmares with their lives on the line. Time. https:\/\/time.com\/5840609\/algorithm-unemployment\/"},{"issue":"12","key":"1572_CR62","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1145\/3458723","volume":"64","author":"T Gebru","year":"2021","unstructured":"Gebru T, Morgenstern JH, Vecchione B, Vaughan JW, Wallach HM, Daum\u00e9 H, Crawford K (2021) Datasheets for datasets. Commun ACM 64(12):86\u201392","journal-title":"Commun ACM"},{"key":"1572_CR63","unstructured":"General Services Administration (2020) Artificial intelligence in federal procurement. https:\/\/www.youtube.com\/watch?app=desktop&v=XJsgbGk8BIw"},{"key":"1572_CR64","unstructured":"Government of Canada (2021) Directive on automated decision making. Modified on January 2021. https:\/\/www.tbs-sct.gc.ca\/pol\/doc-eng.aspx?id=32592"},{"key":"1572_CR65","volume-title":"Discrimination and privacy in the information society studies in applied philosophy, epistemology and rational ethics","author":"S Hajian","year":"2013","unstructured":"Hajian S, Domingo-Ferrer J (2013) Direct and indirect discrimination prevention methods. In: Custers B, Calders T, Schermer B, Zarsky T (eds) Discrimination and privacy in the information society studies in applied philosophy, epistemology and rational ethics, vol 3. Springer, Berlin, Heidelberg"},{"key":"1572_CR66","unstructured":"Heikkila M (2022) A Dutch algorithm scandal serves a warning to Europe. Politico. https:\/\/www.politico.eu\/newsletter\/ai-decoded\/a-dutch-algorithm-scandal-serves-a-warning-to-europe-the-ai-act-wont-save-us-2\/"},{"key":"1572_CR67","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1007\/s43681-020-00008-1","volume":"1","author":"M Hickok","year":"2021","unstructured":"Hickok M (2021) Lessons learned from AI ethics principles for future actions. AI Ethics 1:41\u201347. https:\/\/doi.org\/10.1007\/s43681-020-00008-1","journal-title":"AI Ethics"},{"key":"1572_CR68","doi-asserted-by":"publisher","unstructured":"Hickok M, Dorsey C, O\u2019Brien T, Baur D, Ingram K, Chauhan C, Gamundani A (2022) Case study: the distilling of a biased algorithmic decision system through a business lens. https:\/\/doi.org\/10.2139\/ssrn.4019672. https:\/\/osf.io\/preprints\/socarxiv\/t5dhu\/","DOI":"10.2139\/ssrn.4019672"},{"key":"1572_CR69","unstructured":"Hill K (2020) Wrongfully accused by an algorithm. New York Times. https:\/\/www.nytimes.com\/2020\/06\/24\/technology\/facial-recognition-arrest.html"},{"issue":"4\/5","key":"1572_CR70","doi-asserted-by":"publisher","first-page":"6:1","DOI":"10.1147\/JRD.2019.2942288","volume":"63","author":"M Hind","year":"2019","unstructured":"Hind M, Mehta S, Mojsilovic A, Nair RG, Ramamurthy KN, Olteanu A, Varshney KR (2019) Increasing trust in AI services through supplier\u2019s declarations of conformity. IBM J Res Dev 63(4\/5):6:1-6:13. https:\/\/doi.org\/10.1147\/JRD.2019.2942288","journal-title":"IBM J Res Dev"},{"key":"1572_CR71","unstructured":"Holland S, Hosny A, Newman S, Joseph J, Chmielinski K (2018) The dataset nutrition label: a framework to drive higher data quality standards. ArXiv, abs\/1805.03677"},{"key":"1572_CR72","unstructured":"Howden D, Fotiadis A, Stavinoha L, Holst B (2021) Seeing stones: pandemic reveals Palantir's troubling reach in Europe. The Guardian. https:\/\/www.theguardian.com\/world\/2021\/apr\/02\/seeing-stones-pandemic-reveals-palantirs-troubling-reach-in-europe"},{"key":"1572_CR73","unstructured":"https:\/\/www.gov.uk\/government\/publications\/guidelines-for-ai-procurement\/guidelines-for-ai-procurement"},{"key":"1572_CR74","unstructured":"Information Commissioner\u2019s Office (2020) Explaining decisions made with AI. https:\/\/ico.org.uk\/for-organisations\/guide-to-data-protection\/key-dp-themes\/explaining-decisions-made-with-artificial-intelligence\/"},{"key":"1572_CR75","unstructured":"National Institute of Standards and Technology (2022). Draft AI risk management framework. https:\/\/www.nist.gov\/itl\/ai-risk-management-framework"},{"key":"1572_CR76","unstructured":"National Institute of Standards and Technology (2019). Study evaluates effects of race, age, sex on face recognition software. NIST. https:\/\/www.nist.gov\/news-events\/news\/2019\/12\/nist-study-evaluates-effects-race-age-sex-face-recognition-software"},{"key":"1572_CR77","unstructured":"National Artificial Intelligence Initiative Office (2022) Agency inventories of AI use cases. https:\/\/www.ai.gov\/ai-use-case-inventories\/"},{"key":"1572_CR78","unstructured":"K.W. v. Armstrong, 180, Class action (2016). Casetext. https:\/\/casetext.com\/case\/kw-ex-rel-dw-v-armstrong-5"},{"key":"1572_CR79","doi-asserted-by":"crossref","unstructured":"Kaminski ME and Malgieri G (2019) Algorithmic impact assessments under the GDPR: producing multi-layered explanations. International data privacy law, 2020, forthcoming. U of Colorado Law Legal Studies Research Paper No. 19\u201328","DOI":"10.2139\/ssrn.3456224"},{"key":"1572_CR80","first-page":"125","volume":"54","author":"S Katyal","year":"2019","unstructured":"Katyal S (2019) Private accountability in the age of artificial intelligence. 66 UCLA Law Rev 54:125","journal-title":"66 UCLA Law Rev"},{"issue":"1","key":"1572_CR81","first-page":"23","volume":"43","author":"J Kleinberg","year":"2017","unstructured":"Kleinberg J, Mullainathan S, Raghavan M (2017) Inherent trade-offs in the fair determination of risk scores. Proc Innov Theoret Comput Sci 43(1):23","journal-title":"Proc Innov Theoret Comput Sci"},{"key":"1572_CR82","doi-asserted-by":"crossref","unstructured":"Koulish R (2016) Immigration detention in the risk classification assessment era. Connecticut Public Interest Law Journal, Vol. 16, No.1. https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=2865972#","DOI":"10.1007\/978-3-319-24690-1_1"},{"key":"1572_CR83","unstructured":"Kroll JA, Huey J, Barocas S, Felten EW, Reidenberg JR, Robinson DG, Yu H (2017) Accountable algorithms. University of Pennsylvania Law Review. 165(3). 633 https:\/\/scholarship.law.upenn.edu\/penn_law_review\/vol165\/iss3\/3"},{"key":"1572_CR84","unstructured":"L\u2019Assembl\u00e9e nationale (2016) French digital republic act. https:\/\/www.vie-publique.fr\/eclairage\/20301-loi-republique-numerique-7-octobre-2016-loi-lemaire-quels-changemen"},{"key":"1572_CR85","unstructured":"Lander E, Nelson A (2021). Americans need a bill of rights for an AI-powered world. Wired. https:\/\/www.wired.com\/story\/opinion-bill-of-rights-artificial-intelligence\/"},{"key":"1572_CR86","unstructured":"Laperruque J (2017) Taser\u2019s free body cameras are good for cops, not the people. Wired. https:\/\/www.wired.com\/2017\/04\/tasers-free-body-cameras-good-cops-not-people\/"},{"key":"1572_CR87","unstructured":"Larson J, Mattu S, Kirchner L, Angwin J (2016) How we analyzed the COMPAS recidivism algorithm. ProPublica. https:\/\/www.propublica.org\/article\/how-we-analyzed-the-compas-recidivism-algorithm"},{"key":"1572_CR88","unstructured":"Lecher C (2018) What happens when an algorithm cuts your health care, The Verge. https:\/\/www.theverge.com\/2018\/3\/21\/17144260\/healthcare-medicaid-algorithm-arkansas-cerebral-palsy"},{"key":"1572_CR89","unstructured":"Loomis v. Wisconsin, 38 Supreme Court (2017) Casetext. https:\/\/casetext.com\/case\/state-v-loomis-22"},{"key":"1572_CR90","unstructured":"Lyons K (2021) Amazon\u2019s ring now reportedly partners with more than 2000 US police and fire departments. The Verge. https:\/\/www.theverge.com\/2021\/1\/31\/22258856\/amazon-ring-partners-police-fire-security-privacy-cameras."},{"key":"1572_CR91","unstructured":"Matter of Lederman v. King, 26416. New York Supreme Court (2016). Casetext. https:\/\/casetext.com\/case\/lederman-v-king-1"},{"key":"1572_CR92","doi-asserted-by":"publisher","unstructured":"Metcalf J, Moss E, Watkins EA, Singh R, Elish MC (2021) Algorithmic impact assessments and accountability: the co-construction of impacts. ACM Conference on Fairness, Accountability, and Transparency (FAccT \u201921).735\u2013746. https:\/\/doi.org\/10.1145\/3442188.3445935","DOI":"10.1145\/3442188.3445935"},{"key":"1572_CR93","unstructured":"Metz R (2021) Want your unemployment benefits? You may have to submit to facial recognition first. CNN. https:\/\/www.cnn.com\/2021\/07\/23\/tech\/idme-unemployment-facial-recognition\/index.html"},{"key":"1572_CR94","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1146\/annurev-statistics-042720-125902","volume":"8","author":"S Mitchell","year":"2021","unstructured":"Mitchell S, Potash E, Barocas S, D\u2019Amour A, Lum K (2021) Algorithmic fairness: choices, assumptions, and definitions. Annu Rev Stat Appl 8:1. https:\/\/doi.org\/10.1146\/annurev-statistics-042720-125902","journal-title":"Annu Rev Stat Appl"},{"key":"1572_CR95","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.3877437","author":"E Moss","year":"2021","unstructured":"Moss E, Watkins EA, Singh R, Elish MC, Metcalf J (2021) Assembling accountability: algorithmic impact assessment for the public interest. SSRN J. https:\/\/doi.org\/10.2139\/ssrn.3877437","journal-title":"SSRN J"},{"key":"1572_CR96","doi-asserted-by":"publisher","first-page":"773","DOI":"10.15779\/Z385X25D2W","volume":"34","author":"DK Mulligan","year":"2019","unstructured":"Mulligan DK, Bamberger KA (2019) Procurement as policy: administrative process for machine learning. Berkley Technol Law J 34:773. https:\/\/doi.org\/10.15779\/Z385X25D2W","journal-title":"Berkley Technol Law J"},{"key":"1572_CR97","unstructured":"NAACP Legal Defense Fund (2022) Coalition of civil rights groups sends letter calling for federal and state agencies to end the use of ID.me and facial recognition technology. https:\/\/www.naacpldf.org\/news\/coalition-of-civil-rights-groups-sends-letter-calling-for-federal-and-state-agencies-to-end-the-use-of-id-me-and-facial-recognition-technology\/"},{"key":"1572_CR98","unstructured":"Netherlands Court of Audit (2022) An audit of 9 algorithms used by the Dutch government. https:\/\/english.rekenkamer.nl\/publications\/reports\/2022\/05\/18\/an-audit-of-9-algorithms-used-by-the-dutch-government"},{"key":"1572_CR99","unstructured":"New Zealand Government (2020) Algorithm charter for Aotearoa New Zealand. https:\/\/data.govt.nz\/use-data\/data-ethics\/government-algorithm-transparency-and-accountability"},{"key":"1572_CR100","unstructured":"Newman LH (2019) Internal docs show how ICE gets surveillance help from local cops. Wired. https:\/\/www.wired.com\/story\/ice-license-plate-surveillance-vigilant-solutions"},{"key":"1572_CR101","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1007\/BF02639315","volume":"2","author":"H Nissenbaum","year":"1996","unstructured":"Nissenbaum H (1996) Accountability in a computerized society. Sci Eng Ethics 2:25\u201342. https:\/\/doi.org\/10.1007\/BF02639315","journal-title":"Sci Eng Ethics"},{"issue":"6464","key":"1572_CR102","doi-asserted-by":"publisher","first-page":"447","DOI":"10.1126\/science.aax2342","volume":"366","author":"Z Obermeyer","year":"2019","unstructured":"Obermeyer Z, Powers B, Vogeli C, Mullainathan S (2019) Dissecting racial bias in an algorithm used to manage the health of populations. Science 366(6464):447\u2013453. https:\/\/doi.org\/10.1126\/science.aax2342","journal-title":"Science"},{"key":"1572_CR103","unstructured":"OECD (2019) OECD AI principles. https:\/\/oecd.ai\/en\/ai-principles"},{"key":"1572_CR104","doi-asserted-by":"publisher","DOI":"10.1787\/02682b01-en","volume-title":"Integrating responsible business conduct in public procurement","author":"OECD","year":"2020","unstructured":"OECD (2020) Integrating responsible business conduct in public procurement. OECD Publishing, Paris"},{"key":"1572_CR105","volume-title":"Weapons of math destruction: how big data increases inequality and threatens democracy","author":"C O'Neil","year":"2016","unstructured":"O\u2019Neil C (2016) Weapons of math destruction: how big data increases inequality and threatens democracy. Penguin, London"},{"key":"1572_CR106","unstructured":"Onuoha M (2016) The point of collection. Data and Society. https:\/\/points.datasociety.net\/the-point-of-collection-8ee44ad7c2fa"},{"key":"1572_CR107","unstructured":"Palantir (2020) Form S-1 registration statement. SEC. https:\/\/www.sec.gov\/Archives\/edgar\/data\/1321655\/000119312520230013\/d904406ds1.htm"},{"key":"1572_CR108","first-page":"235","volume":"9","author":"F Pasquale","year":"2011","unstructured":"Pasquale F (2011) Restoring transparency to automated authority. J Telecommun High Technol Law 9:235\u2013256","journal-title":"J Telecommun High Technol Law"},{"key":"1572_CR109","doi-asserted-by":"publisher","DOI":"10.4159\/harvard.9780674736061","volume-title":"The black box society: the secret algorithms that control money and information","author":"F Pasquale","year":"2015","unstructured":"Pasquale F (2015) The black box society: the secret algorithms that control money and information. Harvard University Press, Cambridge, MA and London"},{"key":"1572_CR110","doi-asserted-by":"publisher","unstructured":"Passi S, Barocas S (2019) Problem formulation and fairness. Proc. of the Conference on Fairness, Accountability, and Transparency. 39048. https:\/\/doi.org\/10.1145\/3287560.3287567","DOI":"10.1145\/3287560.3287567"},{"key":"1572_CR111","unstructured":"Priest D (2021) Ring's police problem never went away. Here's what you still need to know. CNET. https:\/\/www.cnet.com\/home\/security\/rings-police-problem-didnt-go-away-it-just-got-more-transparent\/"},{"key":"1572_CR112","unstructured":"Public Oversight of Surveillance Technologies Act. https:\/\/legistar.council.nyc.gov\/LegislationDetail.aspx?ID=3343878&GUID=996ABB2A-9F4C-4A32-B081-D6F24AB954A0"},{"key":"1572_CR113","unstructured":"Ramos G (2022) Ethics of AI and democracy: UNESCO recommendation\u2019s insights. Turkish Policy Quarterly. http:\/\/turkishpolicy.com\/article\/1091\/ethics-of-ai-and-democracy-unesco-recommendations-insights"},{"key":"1572_CR114","unstructured":"Rappeport A, Hill K (2022) IRS to end use of facial recognition for identity verification. The New York Times. https:\/\/www.nytimes.com\/2022\/02\/07\/us\/politics\/irs-idme-facial-recognition.html"},{"key":"1572_CR115","unstructured":"Reisman et al. (2018) Algorithmic impact assessments: a practical framework for public agency accountability. AI Now"},{"key":"1572_CR116","doi-asserted-by":"crossref","unstructured":"Ribeiro M, Singh S, Guestrin C (2016) Why should I trust you?: explaining the predictions of any classifier. In Proc. of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Demonstrations. 97\u2013101, San Diego, California. Association for Computational Linguistics.","DOI":"10.18653\/v1\/N16-3020"},{"key":"1572_CR117","unstructured":"Richardson R, Schultz JM, Crawford K (2019) Dirty data, bad predictions: how civil rights violations impact police data, predictive policing systems and justice. New York University Law Review Online. https:\/\/www.nyulawreview.org\/wp\u2010content\/uploads\/2019\/04\/NYULawReview\u201094\u2010Richardson_etal\u2010FIN.pdf"},{"key":"1572_CR118","unstructured":"Robertson A (2020) ICE rigged its algorithms to keep immigrants in jail, claims lawsuit. The Verge. https:\/\/www.theverge.com\/2020\/3\/3\/21163013\/ice-new-york-risk-assessment-algorithm-rigged-lawsuit-nyclu-jose-velesaca"},{"key":"1572_CR119","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1038\/s42256-019-0048-x","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin C (2019) Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat Mach Intell 1:206\u2013215","journal-title":"Nat Mach Intell"},{"key":"1572_CR120","unstructured":"Ryan-Mosley T (2021) The NYPD used a controversial facial recognition tool. Here\u2019s what you need to know. MIT Technology Review. https:\/\/www.technologyreview.com\/2021\/04\/09\/1022240\/clearview-ai-nypd-emails\/"},{"key":"1572_CR121","unstructured":"Schwartz P (1992) Data processing and government administration: the failure of the American legal response to the computer, 43 Hastings Law Journal. 1321"},{"key":"1572_CR122","unstructured":"Sculley D, Holt G, Golovin D, Davydov E, Phillips T, Ebner D, Chaudhary V, Young M, Crespo J, Dennison D (2015) Hidden technical debt in machine learning systems. NIPS 2503-2511"},{"key":"1572_CR123","unstructured":"Seattle (2017) Washington, surveillance ordinance 123576, http:\/\/seattle.legistar.com\/ViewReport.ashx?M=R&N=Text&GID=393&ID=2849012&GUID=5B7D2F80-A918-4931-9E2E-88E27478A89E&Title=Legislation+Text"},{"key":"1572_CR124","doi-asserted-by":"publisher","unstructured":"Selbst AD, Boyd D, Friedler S, Venkatasubramanian S, Vertesi J (2018) Fairness and abstraction in sociotechnical systems (August 23, 2018). 2019 ACM Conference on Fairness, Accountability, and Transparency (FAT*), 59\u201368. https:\/\/doi.org\/10.1145\/3287560.3287598","DOI":"10.1145\/3287560.3287598"},{"key":"1572_CR125","unstructured":"Shane S, Wakabayashi D (2018) The business of war: google employees protest work for the pentagon. The New York Times. https:\/\/www.nytimes.com\/2018\/04\/04\/technology\/google-letter-ceo-pentagon-project.html"},{"issue":"4","key":"1572_CR126","doi-asserted-by":"publisher","first-page":"541","DOI":"10.1080\/08838151.2020.1843357","volume":"64","author":"D Shin","year":"2020","unstructured":"Shin D (2020) User perceptions of algorithmic decisions in the personalized AI system: perceptual evaluation of fairness, accountability, transparency, and explainability. J Broadcast Electron Media 64(4):541\u2013565. https:\/\/doi.org\/10.1080\/08838151.2020.1843357","journal-title":"J Broadcast Electron Media"},{"key":"1572_CR127","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijhcs.2020.102551","volume":"146","author":"D Shin","year":"2021","unstructured":"Shin D (2021) The effects of explainability and causability on perception, trust, and acceptance: Implications for explainable AI. Int J Hum Comput Stud 146:102551. https:\/\/doi.org\/10.1016\/j.ijhcs.2020.102551","journal-title":"Int J Hum Comput Stud"},{"key":"1572_CR128","doi-asserted-by":"publisher","DOI":"10.1007\/s00146-022-01525-9","author":"D Shin","year":"2022","unstructured":"Shin D, Lim JS, Ahmad N et al (2022) Understanding user sensemaking in fairness and transparency in algorithms: algorithmic sensemaking in over-the-top platform. AI Soc. https:\/\/doi.org\/10.1007\/s00146-022-01525-9","journal-title":"AI Soc"},{"key":"1572_CR129","doi-asserted-by":"publisher","unstructured":"Sloane M, Chowdhury R, Havens JC, Lazovich T, Rincon AL (2021) AI and procurement \u2013 a primer. https:\/\/doi.org\/10.17609\/bxzf-df18","DOI":"10.17609\/bxzf-df18"},{"key":"1572_CR130","unstructured":"Talla V (2019) Documents reveal ICE using driver location data from local police for deportations. ACLU of Northern California. https:\/\/www.aclu.org\/blog\/immigrants-rights\/ice-and-border-patrol-abuses\/documents-reveal-ice-using-driver-location-data"},{"key":"1572_CR131","unstructured":"U.K. Office for Artificial Intelligence (2020) Guidelines for AI procurement. https:\/\/www.gov.uk\/government\/publications\/guidelines-for-ai-procurement\/guidelines-for-ai-procurement"},{"key":"1572_CR132","unstructured":"UK Government Office for Science (2016). Artificial intelligence: opportunities and implications for the future of decision\u2010making. https:\/\/assets.publishing.service.gov.uk\/government\/uploads\/system\/uploads\/attachment_data\/file\/566075\/gs\u201016\u201019\u2010artificial\u2010intelligence\u2010ai\u2010report.pdf"},{"key":"1572_CR133","unstructured":"United States Executive Order 13960 of December 3, 2020. Promoting the use of trustworthy artificial intelligence in the federal government. https:\/\/www.federalregister.gov\/documents\/2020\/12\/08\/2020-27065\/promoting-the-use-of-trustworthy-artificial-intelligence-in-the-federal-government"},{"key":"1572_CR134","unstructured":"United States Federal Acquisition Regulation \u2013 2022-07. https:\/\/www.acquisition.gov\/far\/part-1"},{"key":"1572_CR135","doi-asserted-by":"crossref","unstructured":"Veale M, Kleek MV, Binns R (2018) Fairness and accountability design needs for algorithmic support in high-stakes public sector decision-making. Proc. of the 2018 CHI Conference on Human Factors in Computing Systems.","DOI":"10.31235\/osf.io\/8kvf4"},{"key":"1572_CR136","unstructured":"Velesaca v. Decker (2020) Casetext. https:\/\/casetext.com\/case\/velesaca-v-decker"},{"key":"1572_CR137","doi-asserted-by":"publisher","unstructured":"Verma S, Rubin J (2018) Fairness definitions explained. In Fair-Ware\u201918: IEEE\/ACM International Workshop on Software Fairness. https:\/\/doi.org\/10.1145\/3194770.3194776","DOI":"10.1145\/3194770.3194776"},{"key":"1572_CR138","unstructured":"Wang N, McDonald A, Bateyko D, Tucker E (2022) American dragnet: data-driven deportation in the 21st century, Center on Privacy and Technology at Georgetown Law"},{"key":"1572_CR139","volume-title":"Computer power and human reason: from judgment to calculation","author":"J Weizenbaum","year":"1976","unstructured":"Weizenbaum J (1976) Computer power and human reason: from judgment to calculation. W. H. Freeman & Co"},{"key":"1572_CR140","doi-asserted-by":"publisher","unstructured":"Wieringa M (2020). What to account for when accounting for algorithms: a systematic literature review on algorithmic accountability. Proc.of the 2020 Conference on Fairness, Accountability, and Transparency. 1\u201318. https:\/\/doi.org\/10.1145\/3351095.3372833","DOI":"10.1145\/3351095.3372833"},{"key":"1572_CR141","unstructured":"Winston A (2018) Palantir has secretly been using New Orleans to test its predictive policing technology. The Verge. https:\/\/www.theverge.com\/2018\/2\/27\/17054740\/palantir-predictive-policing-tool-new-orleans-nopd"},{"key":"1572_CR142","doi-asserted-by":"crossref","unstructured":"Young MM, Bullock JB, Lecy JD (2019) Artificial discretion as a tool of governance: a framework for understanding the impact of artificial intelligence on public administration. Perspectives on Public Management and Governance 2(4). https:\/\/academic.oup.com\/ppmg\/article-abstract\/2\/4\/301\/5602198","DOI":"10.1093\/ppmgov\/gvz014"}],"container-title":["AI &amp; SOCIETY"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00146-022-01572-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00146-022-01572-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00146-022-01572-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,9]],"date-time":"2025-04-09T14:03:39Z","timestamp":1744207419000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00146-022-01572-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,10,2]]},"references-count":142,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,6]]}},"alternative-id":["1572"],"URL":"https:\/\/doi.org\/10.1007\/s00146-022-01572-2","relation":{},"ISSN":["0951-5666","1435-5655"],"issn-type":[{"value":"0951-5666","type":"print"},{"value":"1435-5655","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,10,2]]},"assertion":[{"value":"7 June 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 September 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 October 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}