{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T20:14:48Z","timestamp":1776111288159,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":185,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,6,3]],"date-time":"2024-06-03T00:00:00Z","timestamp":1717372800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["IIS2040929"],"award-info":[{"award-number":["IIS2040929"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,6,3]]},"DOI":"10.1145\/3630106.3658910","type":"proceedings-article","created":{"date-parts":[[2024,6,5]],"date-time":"2024-06-05T13:14:21Z","timestamp":1717593261000},"page":"337-358","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":14,"title":["The Fall of an Algorithm: Characterizing the Dynamics Toward Abandonment"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-3180-3582","authenticated-orcid":false,"given":"Nari","family":"Johnson","sequence":"first","affiliation":[{"name":"Carnegie Mellon University, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-4899-5309","authenticated-orcid":false,"given":"Sanika","family":"Moharana","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1850-6459","authenticated-orcid":false,"given":"Christina","family":"Harrington","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3257-2527","authenticated-orcid":false,"given":"Nazanin","family":"Andalibi","sequence":"additional","affiliation":[{"name":"University of Michigan, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3710-4076","authenticated-orcid":false,"given":"Hoda","family":"Heidari","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1499-3045","authenticated-orcid":false,"given":"Motahhare","family":"Eslami","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, USA"}]}],"member":"320","published-online":{"date-parts":[[2024,6,5]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"AlgorithmWatch. Retrieved","year":"2020","unstructured":"AlgorithmWatch 2020. How Dutch activists got an invasive fraud detection algorithm banned. AlgorithmWatch. Retrieved January 16, 2024 from https:\/\/algorithmwatch.org\/en\/syri-netherlands-algorithm\/Email newsletter."},{"key":"e_1_3_2_1_2_1","volume-title":"Retrieved","year":"2020","unstructured":"Logic(s) Magazine 2020. Safe or Just Surveilled?: Tawana Petty on the Fight Against Facial Recognition Surveillance. Logic(s) Magazine. Retrieved January 16, 2024 from https:\/\/logicmag.io\/security\/safe-or-just-surveilled-tawana-petty-on-facial-recognition\/Interview."},{"key":"e_1_3_2_1_3_1","unstructured":"Stop LAPD\u00a0Spying Coalition (Ed.). 2021. The Ghosts of White Supremacy in AI Reform. https:\/\/ainowinstitute.org\/publication\/a-new-ai-lexicon-surveillance"},{"key":"e_1_3_2_1_4_1","volume-title":"Retrieved","author":"Ed Amnesty International","year":"2022","unstructured":"Amnesty International (Ed.). 2022. Ban the Scan. Retrieved January 16, 2024 from https:\/\/banthescan.amnesty.org\/"},{"key":"e_1_3_2_1_5_1","unstructured":"Eticas 2022. The External Audit of the VioGen System. Eticas. https:\/\/eticasfoundation.org\/wp-content\/uploads\/2022\/03\/ETICAS-FND-The-External-Audit-of-the-VioGen-System.pdf"},{"key":"e_1_3_2_1_6_1","unstructured":"2022. Galactica Demo \u2014 galactica.org. https:\/\/galactica.org\/. [Accessed 22-01-2024]."},{"key":"e_1_3_2_1_7_1","volume-title":"Retrieved","author":"Ed The\u00a0Associated","year":"2022","unstructured":"The\u00a0Associated Press (Ed.). 2022. Oregon is dropping an artificial intelligence tool used in child welfare system. Retrieved January 16, 2024 from https:\/\/www.npr.org\/2022\/06\/02\/1102661376\/oregon-drops-artificial-intelligence-child-abuse-cases"},{"key":"e_1_3_2_1_8_1","volume-title":"Retrieved","author":"AI\u00a0Now Institute (Ed.).","year":"2023","unstructured":"AI\u00a0Now Institute (Ed.). 2023. Algorithmic Accountability: Moving Beyond Audits. Retrieved January 16, 2024 from https:\/\/ainowinstitute.org\/publication\/algorithmic-accountability"},{"key":"e_1_3_2_1_9_1","volume-title":"Electronic Frontier Foundation. Retrieved","author":"Foundation Electronic Frontier","year":"2023","unstructured":"Electronic Frontier Foundation 2023. FOIA How To. Electronic Frontier Foundation. Retrieved January 16, 2024 from https:\/\/www.eff.org\/issues\/transparency\/foia-how-to"},{"key":"e_1_3_2_1_10_1","volume-title":"Global AI Legislation Tracker","author":"International Association of Privacy Associates 2023.","year":"2024","unstructured":"International Association of Privacy Associates 2023. Global AI Legislation Tracker. International Association of Privacy Associates. Retrieved January 16, 2024 from https:\/\/iapp.org\/resources\/article\/global-ai-legislation-tracker\/"},{"key":"e_1_3_2_1_11_1","unstructured":"2023. New liver transplant rules yield winners losers as wasted organs reach record high \u2014 washingtonpost.com. https:\/\/www.washingtonpost.com\/business\/2023\/03\/21\/liver-transplants-acuity-circle-policy\/. [Accessed 22-01-2024]."},{"key":"e_1_3_2_1_12_1","volume-title":"Attempts To Revolutionize Mental Health Services. Retrieved","author":"Morris\u00a0Government Duane","year":"2023","unstructured":"Duane Morris\u00a0Government Strategies (Ed.). 2023. Regulating Artificial Intelligence In Mental Health: States\u2019 Attempts To Revolutionize Mental Health Services. Retrieved January 16, 2024 from https:\/\/statecapitallobbyist.com\/artificial-intelligence-ai\/regulating-artificial-intelligence-in-mental-health-services-states-attempts-to-revolutionize-mental-health-services\/"},{"key":"e_1_3_2_1_13_1","volume-title":"Federal Trade Commission. Retrieved","author":"The","year":"2024","unstructured":"The U.S. Federal Trade Commission 2023. Rite Aid Banned from Using AI Facial Recognition After FTC Says Retailer Deployed Technology without Reasonable Safeguards. The U.S. Federal Trade Commission. Retrieved January 16, 2024 from https:\/\/www.ftc.gov\/news-events\/news\/press-releases\/2023\/12\/rite-aid-banned-using-ai-facial-recognition-after-ftc-says-retailer-deployed-technology-withoutPress Release."},{"key":"e_1_3_2_1_14_1","volume-title":"Status of Policing in India Report 2023: Surveillance and the Question of Privacy. Common Cause & Lokniti \u2013 Centre for the Study Developing Societies (CSDS). Retrieved","author":"Cause Common","year":"2024","unstructured":"Common Cause & Lokniti \u2013 Centre for the Study Developing Societies (CSDS) 2023. Status of Policing in India Report 2023: Surveillance and the Question of Privacy. Common Cause & Lokniti \u2013 Centre for the Study Developing Societies (CSDS). Retrieved January 16, 2024 from https:\/\/www.commoncause.in\/wotadmin\/upload\/REPORT_2023.pdf"},{"key":"e_1_3_2_1_15_1","volume-title":"Lucy Parsons Labs. Retrieved","author":"Labs Lucy Parsons","year":"2023","unstructured":"Lucy Parsons Labs 2023. Who We Are. Lucy Parsons Labs. Retrieved January 16, 2024 from https:\/\/lucyparsonslabs.com\/about\/"},{"key":"e_1_3_2_1_16_1","unstructured":"Parag Agrawal and Dantley Davis. 2020. Transparency around image cropping and changes to come. https:\/\/blog.twitter.com\/en_us\/topics\/product\/2020\/transparency-image-cropping"},{"key":"e_1_3_2_1_17_1","volume-title":"Retrieved","author":"Ahmed Sarah","year":"2017","unstructured":"Sarah Ahmed. 2017. No. Retrieved January 16, 2024 from https:\/\/feministkilljoys.com\/2017\/06\/30\/no\/"},{"key":"e_1_3_2_1_18_1","volume-title":"Retrieved","author":"AIAAIC.","year":"2023","unstructured":"AIAAIC. 2023. AI, Algorithmic, and Automation Incidents and Controversies (AIAAIC) Repository. Retrieved January 18, 2024 from https:\/\/www.aiaaic.org\/aiaaic-repository"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300760"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3580970"},{"key":"e_1_3_2_1_21_1","unstructured":"Laboratorio de Inteligencia\u00a0Artificial Aplicada. 2018. Sobre la predicci\u00f3n autom\u00e1tica de embarazos adolescentes."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"crossref","unstructured":"Vijay Arya Rachel K.\u00a0E. Bellamy Pin-Yu Chen Amit Dhurandhar Michael Hind Samuel\u00a0C. Hoffman Stephanie Houde Q.\u00a0Vera Liao Ronny Luss Aleksandra Mojsilovi\u0107 Sami Mourad Pablo Pedemonte Ramya Raghavendra John Richards Prasanna Sattigeri Karthikeyan Shanmugam Moninder Singh Kush\u00a0R. Varshney Dennis Wei and Yunfeng Zhang. 2019. One Explanation Does Not Fit All: A Toolkit and Taxonomy of AI Explainability Techniques. arxiv:1909.03012\u00a0[cs.AI]","DOI":"10.1145\/3351095.3375667"},{"key":"e_1_3_2_1_23_1","volume-title":"Fairness and Machine Learning: Limitations and Opportunities","author":"Barocas Solon","unstructured":"Solon Barocas, Moritz Hardt, and Arvind Narayanan. 2023. Fairness and Machine Learning: Limitations and Opportunities. MIT Press."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/1978942.1979275"},{"key":"e_1_3_2_1_25_1","unstructured":"Rachel K.\u00a0E. Bellamy Kuntal Dey Michael Hind Samuel\u00a0C. Hoffman Stephanie Houde Kalapriya Kannan Pranay Lohia Jacquelyn Martino Sameep Mehta Aleksandra Mojsilovic Seema Nagar Karthikeyan\u00a0Natesan Ramamurthy John Richards Diptikalyan Saha Prasanna Sattigeri Moninder Singh Kush\u00a0R. Varshney and Yunfeng Zhang. 2018. AI Fairness 360: An Extensible Toolkit for Detecting Understanding and Mitigating Unwanted Algorithmic Bias. arxiv:1810.01943\u00a0[cs.AI]"},{"key":"e_1_3_2_1_26_1","unstructured":"Emily Bembeneck Rebecca Nissan and Ziad Obermeyer. 2021. Algorithmic Bias Playbook. https:\/\/www.ftc.gov\/system\/files\/documents\/public_events\/1582978\/algorithmic-bias-playbook.pdf"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1177\/0162243916656059"},{"key":"e_1_3_2_1_28_1","volume-title":"Race after technology: Abolitionist tools for the new Jim code","author":"Benjamin Ruha","unstructured":"Ruha Benjamin. 2019. Race after technology: Abolitionist tools for the new Jim code. Polity."},{"key":"e_1_3_2_1_29_1","volume-title":"Retrieved","author":"Bhuiyan Johana","year":"2021","unstructured":"Johana Bhuiyan. 2021. LAPD ended predictive policing programs amid public outcry. A new effort shares many of their flaws. Retrieved January 16, 2024 from https:\/\/www.theguardian.com\/us-news\/2021\/nov\/07\/lapd-predictive-policing-surveillance-reform"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1191\/1478088706qp063oa"},{"key":"e_1_3_2_1_31_1","unstructured":"Rachel Bukowitz and Tim O\u2019Loughlin. 2022. Governing by Algorithm? Child Protection in Aotearoa New Zealand. https:\/\/anzsog.edu.au\/research-insights-and-resources\/research\/governing-by-algorithm-child-protection-in-aotearoa-new-zealand\/"},{"key":"e_1_3_2_1_32_1","unstructured":"bundesverfassungsgericht. 2023. Bundesverfassungsgericht - Press - Legislation in Hesse and Hamburg regarding automated data analysis for the prevention of criminal acts is unconstitutional. https:\/\/www.bundesverfassungsgericht.de\/SharedDocs\/Pressemitteilungen\/EN\/2023\/bvg23-018.html"},{"key":"e_1_3_2_1_33_1","volume-title":"Conference on fairness, accountability and transparency. PMLR, 77\u201391","author":"Buolamwini Joy","year":"2018","unstructured":"Joy Buolamwini and Timnit Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency. PMLR, 77\u201391."},{"key":"e_1_3_2_1_34_1","volume-title":"Algorithmic Justice League Gender Shades 5th Anniversary Celebration. (2023). https:\/\/www.youtube.com\/watch?v=8JSxbZyivuE Virtual discussion (recorded)","author":"Buolamwini Joy","unstructured":"Joy Buolamwini and Timnit Gebru. 2023. Algorithmic Justice League Gender Shades 5th Anniversary Celebration. (2023). https:\/\/www.youtube.com\/watch?v=8JSxbZyivuE Virtual discussion (recorded)."},{"key":"e_1_3_2_1_35_1","volume-title":"The Berkeley Algorithmic Fairness and Opacity Group Refusal Conference. Retrieved","author":"Burrell Jenna","year":"2020","unstructured":"Jenna Burrell and Deirdre Mulligan. 2020. The Berkeley Algorithmic Fairness and Opacity Group Refusal Conference. Retrieved January 16, 2024 from https:\/\/afog.berkeley.edu\/programs\/the-refusal-conference"},{"key":"e_1_3_2_1_36_1","volume-title":"The Automated Administrative State: A Crisis of Legitimacy. Emory Law Journal","author":"Calo Ryan","year":"2021","unstructured":"Ryan Calo and Danielle\u00a0Keats Citron. 2021. The Automated Administrative State: A Crisis of Legitimacy. Emory Law Journal (2021). https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=3553590"},{"key":"e_1_3_2_1_37_1","unstructured":"Zeph Capo and Janet Bass. 2017. Federal Suit Settlement: End of Value-Added Measures for Teacher Termination in Houston | American Federation of Teachers. https:\/\/www.aft.org\/press-release\/federal-suit-settlement-end-value-added-measures-teacher-termination-houston"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1177\/0163443715584098"},{"key":"e_1_3_2_1_39_1","volume-title":"Retrieved","author":"Chang Cindy","year":"2018","unstructured":"Cindy Chang. 2018. LAPD officials defend predictive policing as activists call for its end. Retrieved January 16, 2024 from https:\/\/www.latimes.com\/local\/lanow\/la-me-lapd-data-policing-20180724-story.html"},{"key":"e_1_3_2_1_40_1","volume-title":"Retrieved","author":"Chayka Kyle","year":"2023","unstructured":"Kyle Chayka. 2023. Rethinking the Luddites in the Age of AI. Retrieved January 16, 2024 from https:\/\/www.newyorker.com\/books\/page-turner\/rethinking-the-luddites-in-the-age-of-ai"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3501831"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3555101"},{"key":"e_1_3_2_1_43_1","unstructured":"Rumman Chowdhury. 2021. Sharing learnings about our image cropping algorithm. https:\/\/blog.twitter.com\/engineering\/en_us\/topics\/insights\/2021\/sharing-learnings-about-our-image-cropping-algorithm"},{"key":"e_1_3_2_1_44_1","volume-title":"Feminist Data Manifest-No. Retrieved","author":"Cifor M.","year":"2024","unstructured":"M. Cifor, P. Garcia, T.L. Cowan, J. Rault, T. Sutherland, A. Chan, J. Rode, A.L. Hoffmann, N. Salehi, and L. Nakamura. 2019. Feminist Data Manifest-No. Retrieved January 16, 2024 from https:\/\/www.manifestno.com\/"},{"key":"e_1_3_2_1_45_1","volume-title":"Retrieved","author":"Conniff Richard","year":"2011","unstructured":"Richard Conniff. 2011. What the Luddites Really Fought Against. Retrieved January 16, 2024 from https:\/\/www.smithsonianmag.com\/history\/what-the-luddites-really-fought-against-264412\/"},{"key":"e_1_3_2_1_46_1","unstructured":"Jeffrey Dastin. [n. d.]. Insight - Amazon scraps secret AI recruiting tool that showed bias against women. https:\/\/www.reuters.com\/article\/us-amazon-com-jobs-automation-insight\/amazon-scraps-secret-ai-recruiting-tool-that-showed-bias-against-women-idUSKCN1MK08G\/. [Accessed 22-01-2024]."},{"key":"e_1_3_2_1_47_1","volume-title":"Retrieved","author":"Dastin Jeffrey","year":"2018","unstructured":"Jeffrey Dastin. 2018. Insight - Amazon scraps secret AI recruiting tool that showed bias against women. Retrieved January 16, 2024 from https:\/\/www.reuters.com\/article\/idUSKCN1MK0AG\/"},{"key":"e_1_3_2_1_48_1","volume-title":"Retrieved","author":"Dastin Jeffrey","year":"2020","unstructured":"Jeffrey Dastin. 2020. Rite Aid deployed facial recognition systems in hundreds of U.S. stores. Retrieved January 16, 2024 from https:\/\/www.reuters.com\/investigates\/special-report\/usa-riteaid-software\/"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1177\/20539517211044808"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376638"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533113"},{"key":"e_1_3_2_1_52_1","volume-title":"Shifting, & Employing Power: A Taxonomy of Responses From Below to Algorithmic Harm. In submission","author":"DeVrio Alicia","year":"2023","unstructured":"Alicia DeVrio, Motahhare Eslami, and Kenneth Holstein. 2023. Building, Shifting, & Employing Power: A Taxonomy of Responses From Below to Algorithmic Harm. In submission (2023)."},{"key":"e_1_3_2_1_53_1","unstructured":"Megan\u00a0Rose Dickey. 2020. Twitter and Zoom\u2019s algorithmic bias issues. https:\/\/techcrunch.com\/2020\/09\/21\/twitter-and-zoom-algorithmic-bias-issues\/"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/2901790.2901861"},{"key":"e_1_3_2_1_55_1","unstructured":"Finale Doshi-Velez and Been Kim. 2017. Towards A Rigorous Science of Interpretable Machine Learning. arxiv:1702.08608\u00a0[stat.ML]"},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/2090236.2090255"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533186"},{"key":"e_1_3_2_1_58_1","volume-title":"Retrieved","author":"Elizalde Elizabeth","year":"2019","unstructured":"Elizabeth Elizalde and Michael Gartland. 2019. Brooklyn tenants in rent-regulated apartments push state to nix landlord\u2019s facial recognition software. Retrieved January 16, 2024 from https:\/\/www.nydailynews.com\/2019\/05\/01\/brooklyn-tenants-in-rent-regulated-apartments-push-state-to-nix-landlords-facial-recognition-software\/"},{"key":"e_1_3_2_1_59_1","volume-title":"Retrieved","author":"Eppink Ritchie","year":"2023","unstructured":"Ritchie Eppink. 2023. Testimony of Ritchie Eppink: AI in Government United States Senate Committee on Homeland Security & Government Affairs. Retrieved January 16, 2024 from https:\/\/www.hsgac.senate.gov\/wp-content\/uploads\/Testimony-Eppink-2023-05-16-1.pdf"},{"key":"e_1_3_2_1_60_1","volume-title":"Retrieved","author":"Erdbrink Thomas","year":"2021","unstructured":"Thomas Erdbrink. 2021. Government in Netherlands Resigns After Benefit Scandal. Retrieved January 16, 2024 from https:\/\/www.nytimes.com\/2021\/01\/15\/world\/europe\/dutch-government-resignation-rutte-netherlands.html"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1609\/icwsm.v11i1.14898"},{"key":"e_1_3_2_1_62_1","unstructured":"Virginia Eubanks. 2018. A response to Allegheny County DHS. https:\/\/virginia-eubanks.com\/2018\/02\/16\/a-response-to-allegheny-county-dhs\/"},{"key":"e_1_3_2_1_63_1","volume-title":"Retrieved","year":"2023","unstructured":"FairFare. 2023. FairFare: Unveiling Ridehail Fairness. Retrieved January 16, 2024 from https:\/\/getfairfare.org\/"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"crossref","unstructured":"Katherine\u00a0B. Forrest. 2021. When Machines Can Be Judge Jury and Executioner: Justice in the Age of Artificial Intelligence. World Scientific.","DOI":"10.1142\/12172"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","unstructured":"Batya Friedman Peter Kahn Alan Borning Ping Zhang and Dennis Galletta. 2006. Value Sensitive Design and Information Systems. https:\/\/doi.org\/10.1007\/978-94-007-7844-3_4","DOI":"10.1007\/978-94-007-7844-3_4"},{"key":"e_1_3_2_1_66_1","volume-title":"Retrieved","author":"Gagne Yasmin","year":"2019","unstructured":"Yasmin Gagne. 2019. How we fought our landlord\u2019s secretive plan for facial recognition\u2014and won. Retrieved January 16, 2024 from https:\/\/www.fastcompany.com\/90431686\/our-landlord-wants-to-install-facial-recognition-in-our-homes-but-were-fighting-back"},{"key":"e_1_3_2_1_67_1","unstructured":"William Gavin. 2022. As privacy concerns arise organizations using facial recognition technology spend on lobbying. https:\/\/www.opensecrets.org\/news\/2022\/03\/as-privacy-concerns-arise-organizations-using-facial-recognition-technology-continue-to-employ-lobbyists\/"},{"key":"e_1_3_2_1_68_1","unstructured":"Albert Gehami and Leila Doty. 2023. When the Rubber Meets the Road: Experience Implementing AI Governance in a Public Agency with the City of San Jos\u00e9. (2023). https:\/\/www.youtube.com\/watch?v=Bif3fwI_d20 ACM FAccT Tutorial."},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3594081"},{"key":"e_1_3_2_1_70_1","volume-title":"Retrieved","author":"Gurley Lauren\u00a0Kaori","year":"2021","unstructured":"Lauren\u00a0Kaori Gurley. 2021. Amazon\u2019s AI Cameras Are Punishing Drivers for Mistakes They Didn\u2019t Make. Retrieved January 16, 2024 from https:\/\/www.vice.com\/en\/article\/88npjv\/amazons-ai-cameras-are-punishing-drivers-for-mistakes-they-didnt-make"},{"key":"e_1_3_2_1_71_1","volume-title":"Retrieved","author":"Hanchett Rebecca","year":"2022","unstructured":"Rebecca Hanchett. 2022. Rhode Island Looks At Banning Facial Recognition By Sports Betting Apps. Retrieved January 16, 2024 from https:\/\/www.gamingtoday.com\/news\/rhode-island-banning-facial-recognition-sports-betting-apps\/"},{"key":"e_1_3_2_1_72_1","unstructured":"Karen Hao. 2020. The two-year fight to stop Amazon from selling face recognition to the police. https:\/\/www.technologyreview.com\/2020\/06\/12\/1003482\/amazon-stopped-selling-police-face-recognition-fight\/"},{"key":"e_1_3_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.7208\/chicago\/9780226315997.001.0001"},{"key":"e_1_3_2_1_74_1","volume-title":"Advances in Neural Information Processing Systems, D.\u00a0Lee, M.\u00a0Sugiyama, U.\u00a0Luxburg, I.\u00a0Guyon, and R.\u00a0Garnett (Eds.). Vol.\u00a029. Curran Associates","author":"Hardt Moritz","year":"2016","unstructured":"Moritz Hardt, Eric Price, Eric Price, and Nati Srebro. 2016. Equality of Opportunity in Supervised Learning. In Advances in Neural Information Processing Systems, D.\u00a0Lee, M.\u00a0Sugiyama, U.\u00a0Luxburg, I.\u00a0Guyon, and R.\u00a0Garnett (Eds.). Vol.\u00a029. Curran Associates, Inc.https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2016\/file\/9d2682367c3935defcb1f9e247a97c0d-Paper.pdf"},{"key":"e_1_3_2_1_75_1","unstructured":"Caroline Haskins. 2019. Dozens of Cities Have Secretly Experimented With Predictive Policing Software. https:\/\/www.vice.com\/en\/article\/d3m7jq\/dozens-of-cities-have-secretly-experimented-with-predictive-policing-software"},{"key":"e_1_3_2_1_76_1","unstructured":"Caroline Haskins. 2021. The NYPD Has Misled The Public About Its Use Of Facial Recognition Tool Clearview AI. https:\/\/www.buzzfeednews.com\/article\/carolinehaskins1\/nypd-has-misled-public-about-clearview-ai-use"},{"key":"e_1_3_2_1_77_1","volume-title":"Retrieved","author":"Hill Kashmir","year":"2020","unstructured":"Kashmir Hill. 2020. The Secretive Company That Might End Privacy as We Know It. Retrieved January 16, 2024 from https:\/\/www.nytimes.com\/2020\/01\/18\/technology\/clearview-privacy-facial-recognition.html"},{"key":"e_1_3_2_1_78_1","volume-title":"Wrongfully Accused by an Algorithm. The New York Times (Jun","author":"Hill Kashmir","year":"2020","unstructured":"Kashmir Hill. 2020. Wrongfully Accused by an Algorithm. The New York Times (Jun 2020). https:\/\/www.nytimes.com\/2020\/06\/24\/technology\/facial-recognition-arrest.html"},{"key":"e_1_3_2_1_79_1","volume-title":"Retrieved","author":"Ho Sally","year":"2022","unstructured":"Sally Ho and Garance Burke. 2022. An algorithm that screens for child neglect raises concerns. Retrieved January 16, 2024 from https:\/\/apnews.com\/article\/child-welfare-algorithm-investigation-9497ee937e0053ad4144a86c68241ef1"},{"key":"e_1_3_2_1_80_1","volume-title":"Child welfare algorithm faces Justice Department scrutiny. https:\/\/apnews.com\/article\/justice-scrutinizes-pittsburgh-child-welfare-ai-tool-4f61f45bfc3245fd2556e886c2da988b, lastaccessed =","author":"Ho Sally","year":"2024","unstructured":"Sally Ho and Garnace Burke. 2023. Child welfare algorithm faces Justice Department scrutiny. https:\/\/apnews.com\/article\/justice-scrutinizes-pittsburgh-child-welfare-ai-tool-4f61f45bfc3245fd2556e886c2da988b, lastaccessed = January 16, 2024,"},{"key":"e_1_3_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290607.3310430"},{"key":"e_1_3_2_1_82_1","volume-title":"A Feminist Theory of Refusal","author":"Honig Bonnie","unstructured":"Bonnie Honig. 2021. A Feminist Theory of Refusal. Harvard University Press."},{"key":"e_1_3_2_1_83_1","volume-title":"Retrieved","author":"Hoover Amanda","year":"2023","unstructured":"Amanda Hoover. 2023. An Eating Disorder Chatbot Is Suspended for Giving Harmful Advice. Retrieved January 16, 2024 from https:\/\/www.wired.com\/story\/tessa-chatbot-suspended\/"},{"key":"e_1_3_2_1_84_1","doi-asserted-by":"publisher","DOI":"10.1145\/2858036.2858470"},{"key":"e_1_3_2_1_85_1","doi-asserted-by":"publisher","DOI":"10.1145\/3462447"},{"key":"e_1_3_2_1_86_1","unstructured":"https:\/\/www.theguardian.com\/profile\/justinmccurry. [n. d.]. South Korean AI chatbot pulled from Facebook after hate speech towards minorities \u2014 theguardian.com. https:\/\/www.theguardian.com\/world\/2021\/jan\/14\/time-to-properly-socialise-hate-speech-ai-chatbot-pulled-from-facebook. [Accessed 22-01-2024]."},{"key":"e_1_3_2_1_87_1","volume-title":"Retrieved","author":"Tech\u00a0Advocacy Hub Benefits","year":"2022","unstructured":"Benefits Tech\u00a0Advocacy Hub. 2022. Arkansas Medicaid Home and Community Based Services Hours Cuts. Retrieved January 16, 2024 from https:\/\/www.btah.org\/case-study\/arkansas-medicaid-home-and-community-based-services-hours-cuts.html"},{"key":"e_1_3_2_1_88_1","volume-title":"Retrieved","author":"Tech\u00a0Advocacy Hub Benefits","year":"2022","unstructured":"Benefits Tech\u00a0Advocacy Hub. 2022. Idaho Medicaid Home and Community Based Services Care Cuts. Retrieved January 16, 2024 from https:\/\/www.btah.org\/case-study\/idaho-medicaid-home-and-community-based-services-care-cuts.html"},{"key":"e_1_3_2_1_89_1","volume-title":"Retrieved","author":"Tech\u00a0Advocacy Hub Benefits","year":"2022","unstructured":"Benefits Tech\u00a0Advocacy Hub. 2022. Missouri Medicaid Home and Community Based Services Eligibility Issues. Retrieved January 16, 2024 from https:\/\/www.btah.org\/case-study\/missouri-medicaid-home-and-community-based-services-eligibility-issues.html"},{"key":"e_1_3_2_1_90_1","volume-title":"Retrieved","author":"Tech\u00a0Advocacy Hub Benefits","year":"2022","unstructured":"Benefits Tech\u00a0Advocacy Hub. 2022. Understanding the Lifecycle of Benefits Technology. Retrieved January 16, 2024 from https:\/\/www.btah.org\/lifecycle.html"},{"key":"e_1_3_2_1_91_1","volume-title":"Predictive Policing","author":"Against The Coalition","year":"2024","unstructured":"The Coalition Against Predictive\u00a0Policing in Pittsburgh. 2020. Predictive Policing in Pittsburgh: A Primer. Retrieved January 16, 2024 from https:\/\/capp-pgh.com\/files\/Primer_v1.pdf"},{"key":"e_1_3_2_1_92_1","volume-title":"Retrieved","author":"Against The Coalition","year":"2020","unstructured":"The Coalition Against Predictive\u00a0Policing in Pittsburgh. 2020. Responding to the \u201cCompletion\u201d of Predictive Policing in Pittsburgh. Retrieved January 16, 2024 from https:\/\/capp-pgh.com\/files\/Metro21%20Counter-Statement.pdf"},{"key":"e_1_3_2_1_93_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445901"},{"key":"e_1_3_2_1_94_1","unstructured":"Pratyusha\u00a0Ria Kalluri William Agnew Myra Cheng Kentrell Owens Luca Soldaini and Abeba Birhane. 2023. The Surveillance AI Pipeline. arxiv:2309.15084\u00a0[cs.CV]"},{"key":"e_1_3_2_1_95_1","volume-title":"It Must Be Abolished. Retrieved","author":"Khan Hamid","year":"2021","unstructured":"Hamid Khan and Pete White. 2021. Police Surveillance Can\u2019t Be Reformed. It Must Be Abolished. Retrieved January 16, 2024 from https:\/\/www.vice.com\/en\/article\/xgzj7n\/police-surveillance-cant-be-reformed-it-must-be-abolished"},{"key":"e_1_3_2_1_96_1","volume-title":"Retrieved","author":"Klar Rebecca","year":"2023","unstructured":"Rebecca Klar. 2023. Exclusive: Meta faces pressure to support independent audit of risk oversight committee. Retrieved January 16, 2024 from https:\/\/news.yahoo.com\/exclusive-meta-faces-pressure-support-164507051.html"},{"key":"e_1_3_2_1_97_1","volume-title":"Retrieved","author":"Kraft Amy","year":"2016","unstructured":"Amy Kraft. 2016. Microsoft shuts down AI chatbot after it turned into a Nazi. Retrieved January 16, 2024 from https:\/\/www.cbsnews.com\/news\/microsoft-shuts-down-ai-chatbot-after-it-turned-into-racist-nazi\/"},{"key":"e_1_3_2_1_98_1","volume-title":"Retrieved","author":"Kwon Junhyup","year":"2021","unstructured":"Junhyup Kwon and Hyeong Yun. 2021. AI Chatbot Shut Down After Learning To Talk Like a Racist Asshole. Retrieved January 16, 2024 from https:\/\/www.vice.com\/en\/article\/akd4g5\/ai-chatbot-shut-down-after-learning-to-talk-like-a-racist-asshole"},{"key":"e_1_3_2_1_99_1","unstructured":"Colin Lecher. 2018. What happens when an algorithm cuts your health care. https:\/\/www.theverge.com\/2018\/3\/21\/17144260\/healthcare-medicaid-algorithm-arkansas-cerebral-palsy"},{"key":"e_1_3_2_1_100_1","unstructured":"Nicol\u00a0Turner Lee Paul Resnick and Genie Barton. 2019. Algorithmic bias detection and mitigation: Best practices and policies to reduce consumer harms. https:\/\/www.brookings.edu\/articles\/algorithmic-bias-detection-and-mitigation-best-practices-and-policies-to-reduce-consumer-harms\/"},{"key":"e_1_3_2_1_101_1","unstructured":"Peter Lee. 2016. Learning from Tay\u2019s introduction. https:\/\/blogs.microsoft.com\/blog\/2016\/03\/25\/learning-tays-introduction\/"},{"key":"e_1_3_2_1_102_1","volume-title":"Retrieved","author":"Liang Calvin","year":"2021","unstructured":"Calvin Liang. 2021. Reflexivity, positionality, and disclosure in HCI. Retrieved January 16, 2024 from https:\/\/medium.com\/@caliang\/reflexivity-positionality-and-disclosure-in-hci-3d95007e9916"},{"key":"e_1_3_2_1_103_1","doi-asserted-by":"crossref","unstructured":"Q.\u00a0Vera Liao and Jennifer\u00a0Wortman Vaughan. 2023. AI Transparency in the Age of LLMs: A Human-Centered Research Roadmap. arxiv:2306.01941\u00a0[cs.HC]","DOI":"10.1162\/99608f92.8036d03b"},{"key":"e_1_3_2_1_104_1","volume-title":"Retrieved","author":"Liban Kevin\u00a0De","year":"2018","unstructured":"Kevin\u00a0De Liban. 2018. Comments re: Notice of Rule-Making for ARChoices Program. Retrieved January 21, 2024 from https:\/\/www.arkleg.state.ar.us\/Home\/FTPDocument?path=%2FAssembly%2FMeeting+Attachments%2F430%2F663%2FHandout+1+-Legal+Aid-Kevin+De+Liban.pdf Email providing public comment on proposed revisions to ARChoices program.."},{"key":"e_1_3_2_1_105_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3593992"},{"key":"e_1_3_2_1_106_1","volume-title":"The Palestine Laboratory: How Israel Exports the Technology of Occupation Around the World","author":"Loewenstein Antony","unstructured":"Antony Loewenstein. 2023. The Palestine Laboratory: How Israel Exports the Technology of Occupation Around the World. Verso. https:\/\/www.versobooks.com\/products\/2684-the-palestine-laboratory"},{"key":"e_1_3_2_1_107_1","volume-title":"Stop LAPD Spying Sues LAPD to Uncover Communications with UCLA Professor Who Founded PredPol","author":"Lomibao Lynn","unstructured":"Lynn Lomibao. 2021. Stop LAPD Spying Sues LAPD to Uncover Communications with UCLA Professor Who Founded PredPol, Inc.https:\/\/stoplapdspying.org\/stop-lapd-spying-sues-lapd-for-communications-with-widely-condemned-ucla-professor-who-founded-predpol-inc\/"},{"key":"e_1_3_2_1_108_1","volume-title":"Retrieved","author":"Ly My","year":"2023","unstructured":"My Ly. 2023. Arkansas DHS agrees to pay $460,000 to settle case over in-home care cuts. Retrieved January 16, 2024 from https:\/\/www.arkansasonline.com\/news\/2023\/aug\/09\/arkansas-dhs-agrees-to-pay-460000-to-settle-case\/"},{"key":"e_1_3_2_1_109_1","unstructured":"Angelica Mari. 2022. S\u00e3o Paulo subway ordered to suspend use of facial recognition. https:\/\/www.zdnet.com\/article\/sao-paulo-subway-ordered-to-suspend-use-of-facial-recognition\/"},{"key":"e_1_3_2_1_110_1","unstructured":"Jesse Marx and Lilly Irani. 2021. Redacted. Taller California. https:\/\/www.printedmatter.org\/catalog\/58378\/"},{"key":"e_1_3_2_1_111_1","doi-asserted-by":"publisher","DOI":"10.14506\/ca31.3.01"},{"key":"e_1_3_2_1_112_1","unstructured":"Sean McGregor. 2020. Preventing Repeated Real World AI Failures by Cataloging Incidents: The AI Incident Database. arxiv:2011.08512\u00a0[cs.CY]"},{"key":"e_1_3_2_1_113_1","unstructured":"James Meadway. 2020. \u201cFuck the Algorithm\u201d: How A-Level Students Have Shown the Future of Protest. https:\/\/novaramedia.com\/2020\/08\/17\/fuck-the-algorithm-how-a-level-students-have-shown-future-of-protest\/"},{"key":"e_1_3_2_1_114_1","volume-title":"Retrieved","author":"Mehrotra Dhruv","year":"2023","unstructured":"Dhruv Mehrotra and Dell Cameron. 2023. The Maker of ShotSpotter Is Buying the World\u2019s Most Infamous Predictive Policing Tech. Retrieved January 16, 2024 from https:\/\/www.wired.com\/story\/soundthinking-geolitica-acquisition-predictive-policing\/"},{"key":"e_1_3_2_1_115_1","volume-title":"Blood in the Machine: The Origins of the Rebellion Against Big Tech. Little","author":"Merchant Brian","unstructured":"Brian Merchant. 2023. Blood in the Machine: The Origins of the Rebellion Against Big Tech. Little, Brown and Company."},{"key":"e_1_3_2_1_116_1","volume-title":"Retrieved","author":"Metcalf Jacob","year":"2023","unstructured":"Jacob Metcalf. 2023. What federal agencies can learn from NYC\u2019s AI Hiring Law. Retrieved January 16, 2024 from https:\/\/thehill.com\/opinion\/technology\/4360523-what-federal-agencies-can-learn-from-new-york-citys-ai-hiring-law\/"},{"key":"e_1_3_2_1_117_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3594092"},{"key":"e_1_3_2_1_118_1","volume-title":"Retrieved","year":"2022","unstructured":"Microsoft. 2022. Microsoft Responsible AI Standard, v2 (General Requirements). Retrieved July 21, 2023 from https:\/\/blogs.microsoft.com\/wp-content\/uploads\/prod\/sites\/5\/2022\/06\/Microsoft-Responsible-AI-Standard-v2-General-Requirements-3.pdf"},{"key":"e_1_3_2_1_119_1","unstructured":"Yeshimabeit Milner. 2020. Abolish Big Data. https:\/\/medium.com\/@YESHICAN\/abolish-big-data-ad0871579a41"},{"key":"e_1_3_2_1_120_1","volume-title":"AI is expensive. A search on Google\u2019s chatbot Bard costs the company 10 times more than a regular one, which could amount to several billion dollars.Retrieved","author":"Mok Aaron","year":"2024","unstructured":"Aaron Mok. 2023. AI is expensive. A search on Google\u2019s chatbot Bard costs the company 10 times more than a regular one, which could amount to several billion dollars.Retrieved January 16, 2024 from https:\/\/www.businessinsider.com\/ai-expensive-google-chatbot-bard-may-cost-company-billions-dollars-2023-2"},{"key":"e_1_3_2_1_121_1","volume-title":"Retrieved","author":"Moss Emanuel","year":"2021","unstructured":"Emanuel Moss, Elizabeth\u00a0Anne Watkins, Ranjit Singh, Madeleine\u00a0Clare Elish, and Jacob Metcalf. 2021. Assembling accountability: algorithmic impact assessment for the public interest. Retrieved July 21, 2023 from https:\/\/datasociety.net\/wp-content\/uploads\/2021\/06\/Assembling-Accountability.pdf"},{"key":"e_1_3_2_1_122_1","volume-title":"Retrieved","year":"2023","unstructured":"Mozilla. 2023. Auditing AI: Announcing the 2023 Mozilla Technology Fund Cohort. Retrieved January 16, 2024 from https:\/\/foundation.mozilla.org\/en\/blog\/auditing-ai-announcing-the-2023-mozilla-technology-fund-cohort\/"},{"key":"e_1_3_2_1_123_1","volume-title":"Algorithms of Oppression: How Search Engines Reinforce Racism","author":"Noble Safiya\u00a0Umoja","unstructured":"Safiya\u00a0Umoja Noble. 2018. Algorithms of Oppression: How Search Engines Reinforce Racism. NYU Press. http:\/\/www.jstor.org\/stable\/j.ctt1pwt9w5"},{"key":"e_1_3_2_1_124_1","volume-title":"Uganda: Rights Concerns Over License Plate Tracking | Human Rights Watch. https:\/\/www.hrw.org\/news\/2023\/11\/14\/uganda-rights-concerns-over-license-plate-tracking","author":"Nyeko Oryem","year":"2023","unstructured":"Oryem Nyeko. 2023. Uganda: Rights Concerns Over License Plate Tracking | Human Rights Watch. https:\/\/www.hrw.org\/news\/2023\/11\/14\/uganda-rights-concerns-over-license-plate-tracking"},{"key":"e_1_3_2_1_125_1","volume-title":"Dissecting racial bias in an algorithm used to manage the health of populations. Science 366, 6464","author":"Obermeyer Ziad","year":"2019","unstructured":"Ziad Obermeyer, Brian Powers, Christine Vogeli, and Sendhil Mullainathan. 2019. Dissecting racial bias in an algorithm used to manage the health of populations. Science 366, 6464 (2019), 447\u2013453."},{"key":"e_1_3_2_1_126_1","volume-title":"Retrieved","author":"Canada The\u00a0Government","year":"2023","unstructured":"The\u00a0Government of Canada. 2023. Algorithmic Impact Assessment tool. Retrieved July 21, 2023 from https:\/\/www.canada.ca\/en\/government\/system\/digital-government\/digital-government-innovations\/responsible-use-ai\/algorithmic-impact-assessment.html"},{"key":"e_1_3_2_1_127_1","volume-title":"Blueprint for an AI Bill of Rights: A Vision for Protecting Our Civil Rights in the Algorithmic Age. Retrieved","author":"Science The White","year":"2023","unstructured":"The White House\u00a0Office of Science and Technology Policy. 2022. Blueprint for an AI Bill of Rights: A Vision for Protecting Our Civil Rights in the Algorithmic Age. Retrieved October 2, 2023 from https:\/\/www.whitehouse.gov\/ostp\/news-updates\/2022\/10\/04\/blueprint-for-an-ai-bill-of-rightsa-vision-for-protecting-our-civil-rights-in-the-algorithmic-age\/"},{"key":"e_1_3_2_1_128_1","unstructured":"UN OHCHR. 2022. OHCHR Assessment of Human Rights Concerns in the Xinjiang Uyghur Autonomous Region People\u2019s Republic of China."},{"key":"e_1_3_2_1_129_1","unstructured":"Victor Ojewale Ryan Steed Briana Vecchione Abeba Birhane and Inioluwa\u00a0Deborah Raji. 2024. Towards AI Accountability Infrastructure: Gaps and Opportunities in AI Audit Tooling. arxiv:2402.17861\u00a0[cs.CY]"},{"key":"e_1_3_2_1_130_1","doi-asserted-by":"publisher","DOI":"10.1145\/3527152"},{"key":"e_1_3_2_1_131_1","unstructured":"Phil Pennington. 2023. Facial recognition: Officials yet to meet obligation to seek views of M\u0101ori - documents. https:\/\/www.rnz.co.nz\/news\/national\/501761\/facial-recognition-officials-yet-to-meet-obligation-to-seek-views-of-maori-documents"},{"key":"e_1_3_2_1_132_1","unstructured":"Nick Perry. 2023. New Zealand debates whether ethnicity should be a factor for surgery waitlists. https:\/\/apnews.com\/article\/new-zealand-surgery-ethnicity-algorithm-maori-1b44026f2661772a7eb3bd4444619446"},{"key":"e_1_3_2_1_133_1","doi-asserted-by":"publisher","DOI":"10.1145\/3494672"},{"key":"e_1_3_2_1_134_1","volume-title":"Lawsuit claims UnitedHealth AI wrongfully denies elderly extended care","author":"Pierson Brendan","year":"2023","unstructured":"Brendan Pierson and Brendan Pierson. 2023. Lawsuit claims UnitedHealth AI wrongfully denies elderly extended care. Reuters (Nov 2023). https:\/\/www.reuters.com\/legal\/lawsuit-claims-unitedhealth-ai-wrongfully-denies-elderly-extended-care-2023-11-14\/"},{"key":"e_1_3_2_1_135_1","volume-title":"Retrieved","author":"Porter Jon","year":"2020","unstructured":"Jon Porter. 2020. UK ditches exam results generated by biased algorithm after student protests. Retrieved January 16, 2024 from https:\/\/www.theverge.com\/2020\/8\/17\/21372045\/uk-a-level-results-algorithm-biased-coronavirus-covid-19-pandemic-university-applications"},{"key":"e_1_3_2_1_136_1","doi-asserted-by":"publisher","DOI":"10.3390\/su14106055"},{"key":"e_1_3_2_1_137_1","doi-asserted-by":"publisher","DOI":"10.1145\/3571151"},{"key":"e_1_3_2_1_138_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533158"},{"key":"e_1_3_2_1_139_1","volume-title":"Retrieved","author":"Reisman Dillon","year":"2018","unstructured":"Dillon Reisman, Jason Schultz, Kate Crawford, and Meredith Whittaker. 2018. Algorithmic Impact Assessments: A Practical Framework for Public Agency. Retrieved July 21, 2023 from https:\/\/www.nist.gov\/system\/files\/documents\/2021\/10\/04\/aiareport2018.pdf"},{"key":"e_1_3_2_1_140_1","doi-asserted-by":"crossref","unstructured":"Marco\u00a0Tulio Ribeiro Sameer Singh and Carlos Guestrin. 2016. \"Why Should I Trust You?\": Explaining the Predictions of Any Classifier. arxiv:1602.04938\u00a0[cs.LG]","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_141_1","doi-asserted-by":"publisher","DOI":"10.1145\/3579543"},{"key":"e_1_3_2_1_142_1","volume-title":"Retrieved","author":"Ross Casey","year":"2021","unstructured":"Casey Ross. 2021. Epic\u2019s AI algorithms, shielded from scrutiny by a corporate firewall, are delivering inaccurate information on seriously ill patients. Retrieved January 21, 2024 from https:\/\/www.statnews.com\/2021\/07\/26\/epic-hospital-algorithms-sepsis-investigation\/"},{"key":"e_1_3_2_1_143_1","volume-title":"Retrieved","author":"Ryan-Mosley Tate","year":"2023","unstructured":"Tate Ryan-Mosley. 2023. An algorithm intended to reduce poverty might disqualify people in need. Retrieved January 16, 2024 from https:\/\/www.technologyreview.com\/2023\/06\/13\/1074551\/an-algorithm-intended-to-reduce-poverty-in-jordan-disqualifies-people-in-need\/"},{"key":"e_1_3_2_1_144_1","unstructured":"Tate Ryan-Mosley. 2023. How face recognition rules in the US got stuck in political gridlock. https:\/\/www.technologyreview.com\/2023\/07\/24\/1076668\/how-face-recognition-rules-in-the-us-got-stuck-in-political-gridlock\/"},{"key":"e_1_3_2_1_145_1","doi-asserted-by":"publisher","DOI":"10.1111\/1467-9566.12900"},{"key":"e_1_3_2_1_146_1","volume-title":"Retrieved","author":"Scott Joey","year":"2023","unstructured":"Joey Scott. 2023. LAPD Is Using Israeli Surveillance Software That Can Track Your Phone and Social Media. Retrieved January 16, 2024 from https:\/\/knock-la.com\/lapd-is-using-israeli-surveillance-software-that-can-track-your-phone-and-social-media\/#:\u00a0:text=During%20a%202014%20trip%20to,would%20be%20using%20all%20three"},{"key":"e_1_3_2_1_147_1","doi-asserted-by":"publisher","DOI":"10.1515\/9780691190600-028"},{"key":"e_1_3_2_1_148_1","first-page":"1085","article-title":"The Intuitive Appeal of Explainable Machines","volume":"87","author":"Selbst D.","year":"2018","unstructured":"Andrew\u00a0D. Selbst and Solon Barocas. 2018. The Intuitive Appeal of Explainable Machines. Fordham Law Review 87 (2018), 1085. https:\/\/api.semanticscholar.org\/CorpusID:59548063","journal-title":"Fordham Law Review"},{"key":"e_1_3_2_1_149_1","volume-title":"Deconstructing Design Decisions: Why Courts Must Interrogate Machine Learning and Other Technologies. Ohio State Law Journal 85","author":"Selbst D.","year":"2024","unstructured":"Andrew\u00a0D. Selbst, Suresh Venkatasubramanian, and I.\u00a0Elizabeth Kumar. 2024. Deconstructing Design Decisions: Why Courts Must Interrogate Machine Learning and Other Technologies. Ohio State Law Journal 85 (2024). https:\/\/ssrn.com\/abstract=4564304"},{"key":"e_1_3_2_1_150_1","doi-asserted-by":"publisher","DOI":"10.1145\/3600211.3604673"},{"key":"e_1_3_2_1_151_1","doi-asserted-by":"publisher","DOI":"10.1145\/3479577"},{"key":"e_1_3_2_1_152_1","volume-title":"On Ethnographic Refusal: Indigeneity, \u2018Voice","author":"Simpson Audra","year":"2007","unstructured":"Audra Simpson. 2007. On Ethnographic Refusal: Indigeneity, \u2018Voice\u2019 and Colonial Citizenship. Junctures (2007)."},{"key":"e_1_3_2_1_153_1","volume-title":"Here\u2019s What We Learned. Retrieved","author":"Sisitzky Michael","year":"2021","unstructured":"Michael Sisitzky and Ben Schaefer. 2021. The NYPD Published its Arsenal of Surveillance Tech. Here\u2019s What We Learned. Retrieved January 16, 2024 from https:\/\/www.nyclu.org\/en\/news\/nypd-published-its-arsenal-surveillance-tech-heres-what-we-learned"},{"key":"e_1_3_2_1_154_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220046"},{"key":"e_1_3_2_1_155_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533177"},{"key":"e_1_3_2_1_156_1","unstructured":"Varshini Subhash Zixi Chen Marton Havasi Weiwei Pan and Finale Doshi-Velez. 2022. What Makes a Good Explanation?: A Harmonized View of Properties of Explanations. In Progress and Challenges in Building Trustworthy Embodied AI. https:\/\/openreview.net\/forum?id=YDyLZWwpBK2"},{"key":"e_1_3_2_1_157_1","doi-asserted-by":"publisher","DOI":"10.1145\/3465416.3483305"},{"key":"e_1_3_2_1_158_1","volume-title":"Retrieved","author":"Regional Justice\u00a0Information System The Automated","year":"2021","unstructured":"The Automated Regional Justice\u00a0Information System. 2021. TACIDS: Tactical Identification System Using Facial Recognition. Retrieved January 16, 2024 from https:\/\/voiceofsandiego.org\/wp-content\/uploads\/2021\/04\/TACIDS-Final-Report-FINAL.pdf"},{"key":"e_1_3_2_1_159_1","volume-title":"Native American DNA: Tribal Belonging and the False Promise of Genetic Science","author":"TallBear Kim","unstructured":"Kim TallBear. 2013. Native American DNA: Tribal Belonging and the False Promise of Genetic Science. U of Minnesota Press."},{"key":"e_1_3_2_1_160_1","volume-title":"Automated Neglect: How the World Bank\u2019s Push to Allocate Cash Assistance Using Algorithms Threatens Rights. Human Rights Watch. https:\/\/books.google.com\/books?id=rNMG0AEACAAJ","author":"Toh A.","year":"2023","unstructured":"A. Toh and Human\u00a0Rights Watch. 2023. Automated Neglect: How the World Bank\u2019s Push to Allocate Cash Assistance Using Algorithms Threatens Rights. Human Rights Watch. https:\/\/books.google.com\/books?id=rNMG0AEACAAJ"},{"key":"e_1_3_2_1_161_1","volume-title":"Retrieved","author":"Vaithianathan Rhema","year":"2019","unstructured":"Rhema Vaithianathan, Emily Kulick, Emily Putnam-Hornstein, and Diana\u00a0Benavides Prado. 2019. Allegheny Family Screening Tool: Methodology, Version 2. Retrieved January 21, 2024 from https:\/\/www.alleghenycountyanalytics.us\/wp-content\/uploads\/2019\/05\/Methodology-V2-from-16-ACDHS-26_PredictiveRisk_Package_050119_FINAL-7.pdf"},{"key":"e_1_3_2_1_162_1","volume-title":"Sistema Alerta Ni\u00f1ez y la predicci\u00f3n del riesgo de vulneraci\u00f3n de derechos de la infancia. Derechos Digitales","author":"Valderrama Mat\u00edas","year":"2021","unstructured":"Mat\u00edas Valderrama. 2021. Sistema Alerta Ni\u00f1ez y la predicci\u00f3n del riesgo de vulneraci\u00f3n de derechos de la infancia. Derechos Digitales (2021)."},{"key":"e_1_3_2_1_163_1","volume-title":"Landmark judgment from The Netherlands on digital welfare states and human rights. Open Global Rights","author":"van Veen C","year":"2020","unstructured":"C van Veen. 2020. Landmark judgment from The Netherlands on digital welfare states and human rights. Open Global Rights (2020)."},{"key":"e_1_3_2_1_164_1","volume-title":"Not My A.I.: Towards Critical Feminist Frameworks to Resist Oppressive A.I. Systems","author":"Varon Joana","year":"2022","unstructured":"Joana Varon and Paz Pe\u00f1a. 2022. Not My A.I.: Towards Critical Feminist Frameworks to Resist Oppressive A.I. Systems. Carr Center Discussion Paper Series (2022). https:\/\/carrcenter.hks.harvard.edu\/publications\/notmyai"},{"key":"e_1_3_2_1_165_1","unstructured":"James Vincent. [n. d.]. Twitter taught Microsoft\u2019s AI chatbot to be a racist asshole in less than a day \u2014 theverge.com. https:\/\/www.theverge.com\/2016\/3\/24\/11297050\/tay-microsoft-chatbot-racist. [Accessed 22-01-2024]."},{"key":"e_1_3_2_1_166_1","volume-title":"One Company\u2019s Algorithm Could Be Why.Retrieved","author":"Vogell Heather","year":"2022","unstructured":"Heather Vogell, Haru Coryne, and Ryan Little. 2022. Rent Going Up? One Company\u2019s Algorithm Could Be Why.Retrieved January 16, 2024 from https:\/\/www.propublica.org\/article\/yieldstar-rent-increase-realpage-rent"},{"key":"e_1_3_2_1_167_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3594030"},{"key":"e_1_3_2_1_168_1","volume-title":"Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics(Proceedings of Machine Learning Research, Vol.\u00a038)","author":"Wang Fulton","year":"2015","unstructured":"Fulton Wang and Cynthia Rudin. 2015. Falling Rule Lists. In Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics(Proceedings of Machine Learning Research, Vol.\u00a038), Guy Lebanon and S.\u00a0V.\u00a0N. Vishwanathan (Eds.). PMLR, San Diego, California, USA, 1013\u20131022. https:\/\/proceedings.mlr.press\/v38\/wang15a.html"},{"key":"e_1_3_2_1_169_1","volume-title":"Retrieved","author":"Wang Nina","year":"2022","unstructured":"Nina Wang, Allison McDonald, Daniel Bateyko, and Emily Tucker. 2022. American Dragnet: Data-Driven Deportation in the 21st Century. (2022). Retrieved January 16, 2024 from https:\/\/americandragnet.org\/"},{"key":"e_1_3_2_1_170_1","volume-title":"Warn of De-Facto Price Setting and Collusion After Senate Investigation. Retrieved","author":"Warren Elizabeth","year":"2023","unstructured":"Elizabeth Warren. 2023. Warren, Lawmakers Urge Justice Department to Review YieldStar, Warn of De-Facto Price Setting and Collusion After Senate Investigation. Retrieved January 21, 2024 from https:\/\/www.warren.senate.gov\/oversight\/letters\/warren-lawmakers-urge-justice-department-to-review-yieldstar-warn-of-de-facto-price-setting-and-collusion-after-senate-investigation Press Release."},{"key":"e_1_3_2_1_171_1","volume-title":"Retrieved","author":"Watch Human\u00a0Rights","year":"2023","unstructured":"Human\u00a0Rights Watch. 2023. Automated Neglect: How The World Bank\u2019s Push to Allocate Cash Assistance Using Algorithms Threatens Rights. Retrieved January 16, 2024 from https:\/\/www.hrw.org\/report\/2023\/06\/13\/automated-neglect\/how-world-banks-push-allocate-cash-assistance-using-algorithms"},{"key":"e_1_3_2_1_172_1","volume-title":"Fairlearn: Assessing and Improving Fairness of AI Systems. arxiv:2303.16626\u00a0[cs.LG]","author":"Weerts Hilde","year":"2023","unstructured":"Hilde Weerts, Miroslav Dud\u00edk, Richard Edgar, Adrin Jalali, Roman Lutz, and Michael Madaio. 2023. Fairlearn: Assessing and Improving Fairness of AI Systems. arxiv:2303.16626\u00a0[cs.LG]"},{"key":"e_1_3_2_1_173_1","unstructured":"Emma Weil and Elizabeth Edwards. 2023. Using Technical Skills to Fight Actual Public Benefits Cuts and Austerity Policies with the Benefits Tech Advocacy Hub. (2023). https:\/\/www.youtube.com\/watch?v=ZELeRPx74PE ACM FAccT Tutorial."},{"key":"e_1_3_2_1_174_1","volume-title":"Retrieved","author":"Wells Kate","year":"2023","unstructured":"Kate Wells. 2023. An eating disorders chatbot offered dieting advice, raising fears about AI in health. Retrieved January 16, 2024 from https:\/\/www.npr.org\/sections\/health-shots\/2023\/06\/08\/1180838096\/an-eating-disorders-chatbot-offered-dieting-advice-raising-fears-about-ai-in-hea"},{"key":"e_1_3_2_1_175_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445314"},{"key":"e_1_3_2_1_176_1","volume-title":"Retrieved","author":"Whittaker Zack","year":"2019","unstructured":"Zack Whittaker. 2019. Amazon shareholders reject facial recognition sale ban to governments. Retrieved January 16, 2024 from https:\/\/techcrunch.com\/2019\/05\/22\/amazon-reject-facial-recognition-proposals\/"},{"key":"e_1_3_2_1_177_1","volume-title":"Retrieved","author":"Wilkinson Christopher","year":"2023","unstructured":"Christopher Wilkinson and Dorthy Lukens. 2023. The Growing Regulation of AI-Based Employment Decision Tools. Retrieved January 16, 2024 from https:\/\/www.perkinscoie.com\/en\/news-insights\/the-growing-regulation-of-ai-based-employment-decision-tools.html"},{"key":"e_1_3_2_1_178_1","volume-title":"Retrieved","author":"Williams Timothy","year":"2015","unstructured":"Timothy Williams. 2015. Facial Recognition Software Moves From Overseas Wars to Local Police. Retrieved January 16, 2024 from https:\/\/www.nytimes.com\/2015\/08\/13\/us\/facial-recognition-software-moves-from-overseas-wars-to-local-police.html?_r=1"},{"key":"e_1_3_2_1_179_1","volume-title":"Retrieved","author":"Wilson Tom","year":"2019","unstructured":"Tom Wilson and Madhumita Murgia. 2019. Uganda confirms use of Huawei facial recognition cameras. Retrieved January 16, 2024 from https:\/\/www.ft.com\/content\/e20580de-c35f-11e9-a8e9-296ca66511c9"},{"key":"e_1_3_2_1_180_1","doi-asserted-by":"publisher","DOI":"10.1001\/jamainternmed.2021.2626"},{"key":"e_1_3_2_1_181_1","volume-title":"Retrieved","author":"Wu Sarah","year":"2019","unstructured":"Sarah Wu. 2019. Somerville City Council passes facial recognition ban. Retrieved January 16, 2024 from https:\/\/www.bostonglobe.com\/metro\/2019\/06\/27\/somerville-city-council-passes-facial-recognition-ban\/"},{"key":"e_1_3_2_1_182_1","volume-title":"Responses After Firing Human Staff. Retrieved","author":"Xiang Chloe","year":"2023","unstructured":"Chloe Xiang. 2023. Eating Disorder Helpline Disables Chatbot for \u2019Harmful\u2019 Responses After Firing Human Staff. Retrieved January 16, 2024 from https:\/\/www.vice.com\/en\/article\/qjvk97\/eating-disorder-helpline-disables-chatbot-for-harmful-responses-after-firing-human-staff"},{"key":"e_1_3_2_1_183_1","doi-asserted-by":"publisher","DOI":"10.1145\/3479594"},{"key":"e_1_3_2_1_184_1","doi-asserted-by":"publisher","DOI":"10.1145\/3563657.3596058"},{"key":"e_1_3_2_1_185_1","doi-asserted-by":"publisher","DOI":"10.1145\/3630107"}],"event":{"name":"FAccT '24: The 2024 ACM Conference on Fairness, Accountability, and Transparency","location":"Rio de Janeiro Brazil","acronym":"FAccT '24"},"container-title":["The 2024 ACM Conference on Fairness, Accountability, and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3630106.3658910","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3630106.3658910","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T22:50:56Z","timestamp":1750287056000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3630106.3658910"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,3]]},"references-count":185,"alternative-id":["10.1145\/3630106.3658910","10.1145\/3630106"],"URL":"https:\/\/doi.org\/10.1145\/3630106.3658910","relation":{},"subject":[],"published":{"date-parts":[[2024,6,3]]},"assertion":[{"value":"2024-06-05","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}