{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,25]],"date-time":"2026-04-25T21:35:26Z","timestamp":1777152926208,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":120,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T00:00:00Z","timestamp":1655683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["1928586"],"award-info":[{"award-number":["1928586"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,21]]},"DOI":"10.1145\/3531146.3533186","type":"proceedings-article","created":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T14:27:10Z","timestamp":1655735230000},"page":"1305-1317","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":33,"title":["The Algorithmic Imprint"],"prefix":"10.1145","author":[{"given":"Upol","family":"Ehsan","sequence":"first","affiliation":[{"name":"Georgia Institute of Technology, USA"}]},{"given":"Ranjit","family":"Singh","sequence":"additional","affiliation":[{"name":"AI on the Ground Initiative, Data &amp; Society Research Institute, USA"}]},{"given":"Jacob","family":"Metcalf","sequence":"additional","affiliation":[{"name":"AI on the Ground Initiative, Data &amp; Society Research Institute, USA"}]},{"given":"Mark","family":"Riedl","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology, USA"}]}],"member":"320","published-online":{"date-parts":[[2022,6,20]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372871"},{"key":"e_1_3_2_1_2_1","unstructured":"Ada Lovelace Institute. 2020. Inspecting Algorithms in Social Media Platforms. Retrieved from https:\/\/www.adalovelaceinstitute.org\/algorithms-in-social-media-realistic-routes-to-regulatory-inspection\/"},{"key":"e_1_3_2_1_3_1","volume-title":"Condemns Racially Biased Surveillance. NPR. Retrieved","author":"Allyn Bobby","year":"2022","unstructured":"Bobby Allyn. 2020. IBM Abandons Facial Recognition Products, Condemns Racially Biased Surveillance. NPR. Retrieved January 4, 2022 from https:\/\/www.npr.org\/2020\/06\/09\/873298837\/ibm-abandons-facial-recognition-products-condemns-racially-biased-surveillance"},{"key":"e_1_3_2_1_4_1","volume-title":"The Guardian. Retrieved","author":"Amoore Louise","year":"2020","unstructured":"Louise Amoore. 2020. Why \u201cDitch the algorithm\u201d is the future of political protest. The Guardian. Retrieved January 14, 2022 from https:\/\/www.theguardian.com\/commentisfree\/2020\/aug\/19\/ditch-the-algorithm-generation-students-a-levels-politics"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1177\/0162243920912824"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"J Aronson. 1994. A pragmatic view of thematic analysis: the qualitative report 2 (1) Spring.","DOI":"10.46743\/2160-3715\/1995.2069"},{"key":"e_1_3_2_1_7_1","volume-title":"The Oxford Handbook of Ethics of AI, Markus D","author":"Arun Chinmayi","unstructured":"Chinmayi Arun. 2020. AI and the Global South: Designing for Other Worlds. In The Oxford Handbook of Ethics of AI, Markus D. Dubber, Frank Pasquale and Sunit Das (eds.). Oxford University Press, New York, 589\u2013606."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1080\/14626268.2013.767276"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.2044-8279.1968.tb02005.x"},{"key":"e_1_3_2_1_10_1","unstructured":"Solon Barocas Moritz Hardt and Arvind Narayanan. 2019. Fairness and Machine Learning: Limitations and Opportunities. Retrieved from http:\/\/www.fairmlbook.org"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445922"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3173951"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patter.2021.100205"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"Virginia Braun and Victoria Clarke. 2006. Using thematic analysis in psychology. Qualitative research in psychology 3 2: 77\u2013101.","DOI":"10.1191\/1478088706qp063oa"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1080\/1369118X.2016.1154086"},{"key":"e_1_3_2_1_16_1","volume-title":"Response: Racial and Gender bias in Amazon Rekognition \u2014 Commercial AI System for Analyzing Faces. Medium. Retrieved","author":"Buolamwini Joy","year":"2019","unstructured":"Joy Buolamwini. 2019. Response: Racial and Gender bias in Amazon Rekognition \u2014 Commercial AI System for Analyzing Faces. Medium. Retrieved October 5, 2020 from https:\/\/medium.com\/@Joy.Buolamwini\/response-racial-and-gender-bias-in-amazon-rekognition-commercial-ai-system-for-analyzing-faces-a289222eeced"},{"key":"e_1_3_2_1_17_1","volume-title":"Proceedings of Machine Learning Research, 1\u201315","author":"Buolamwini Joy","year":"2018","unstructured":"Joy Buolamwini and Timnit Gebru. 2018. Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification. In Proceedings of Machine Learning Research, 1\u201315. Retrieved from http:\/\/proceedings.mlr.press\/v81\/buolamwini18a.html"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359240"},{"key":"e_1_3_2_1_19_1","unstructured":"Wendy Hui Kyong Chun. 2011. Programmed visions: Software and memory. mit Press."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3322276.3323690"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1080\/13645579.2019.1649832"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1177\/1527476418796632"},{"key":"e_1_3_2_1_23_1","first-page":"1443","article-title":"AI governance: a research agenda. Governance of AI Program, Future of Humanity Institute, University of Oxford: Oxford","volume":"1442","author":"Dafoe Allan","year":"2018","unstructured":"Allan Dafoe. 2018. AI governance: a research agenda. Governance of AI Program, Future of Humanity Institute, University of Oxford: Oxford, UK 1442: 1443.","journal-title":"UK"},{"key":"e_1_3_2_1_24_1","volume-title":"Divining a Digital Future: Mess and Mythology in Ubiquitous Computing","author":"Dourish Paul","unstructured":"Paul Dourish and Genevieve Bell. 2011. Divining a Digital Future: Mess and Mythology in Ubiquitous Computing. MIT Press, Cambridge."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/2370216.2370238"},{"key":"e_1_3_2_1_26_1","volume-title":"What Do You Need to Know? with Deb Raji. Retrieved","author":"Doyle-Burke Dylan","year":"2020","unstructured":"Dylan Doyle-Burke and Jessie Smith. IBM, Microsoft, and Amazon Disavow Facial Recognition Technology: What Do You Need to Know? with Deb Raji. Retrieved October 6, 2020 from https:\/\/radicalai.podbean.com\/e\/ibm-microsoft-and-amazon-disavow-facial-recognition-technology-what-do-you-need-to-know-with-deb-raji\/"},{"key":"e_1_3_2_1_27_1","volume-title":"Infrastructure and Modernity: Force, Time, and Social Organization in the History of Sociotechnical Systems","author":"Edwards Paul N","unstructured":"Paul N Edwards. 2003. Infrastructure and Modernity: Force, Time, and Social Organization in the History of Sociotechnical Systems. In Modernity and Technology, Thomas J Misa, Philip Brey and Andrew Feenberg (eds.). MIT Press, Cambridge, MA, 185\u2013225."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445188"},{"key":"e_1_3_2_1_29_1","volume-title":"Proceedings of Emerging Perspectives in Human-Centered Machine Learning: A Workshop at The ACM CHI Conference on Human Factors in Computing Systems.","author":"Ehsan Upol","year":"2019","unstructured":"Upol Ehsan and Mark O Riedl. 2019. On Design and Evaluation of Human-centered Explainable AI systems. In Proceedings of Emerging Perspectives in Human-Centered Machine Learning: A Workshop at The ACM CHI Conference on Human Factors in Computing Systems."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-60117-1_33"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3301275.3302316"},{"key":"e_1_3_2_1_32_1","unstructured":"Sl Eiin. Board of Intermediate and Secondary Education Dhaka. 1. Retrieved from https:\/\/dhakaeducationboard.gov.bd\/data\/20200813101513549205.pdf"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1548-1425.2010.01265.x"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/2702123.2702556"},{"key":"e_1_3_2_1_35_1","volume-title":"Automating inequality: How high-tech tools profile, police, and punish the poor","author":"Eubanks Virginia","unstructured":"Virginia Eubanks. 2018. Automating inequality: How high-tech tools profile, police, and punish the poor. St. Martin's Press."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","unstructured":"Sina Fazelpour Zachary C. Lipton and David Danks. 2021. Algorithmic Fairness and the Situated Dynamics of Justice. Canadian Journal of Philosophy: 1\u201317. https:\/\/doi.org\/10.1017\/can.2021.24","DOI":"10.1017\/can.2021.24"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"crossref","unstructured":"Batya Friedman Peter H Kahn and Alan Borning. 2008. Value sensitive design and information systems. The handbook of information and computer ethics: 69\u2013101.","DOI":"10.1002\/9780470281819.ch4"},{"key":"e_1_3_2_1_38_1","volume-title":"David Maines (ed.)","author":"Fujimura Joan H","unstructured":"Joan H Fujimura. 1991. On Methods, Ontologies, and Representation in the Sociology of Science: Where Do We Stand? In Social Organization and Social Process: Essays in Honor of Anselm Strauss, David Maines (ed.). Aldine de Gruyter, Hawthorne, NY, 207\u2013248."},{"key":"e_1_3_2_1_39_1","volume-title":"Proceedings of the 5th Workshop on Fairness, Accountability, and Transparency in Machine Learning. Retrieved","author":"Gebru Timnit","year":"2018","unstructured":"Timnit Gebru, Jamie Morgenstern, Briana Vecchione, Jennifer Wortman Vaughan, Hanna Wallach, Hal Daum\u00e9 III, and Kate Crawford. 2018. Datasheets for Datasets. In Proceedings of the 5th Workshop on Fairness, Accountability, and Transparency in Machine Learning. Retrieved July 6, 2020 from http:\/\/arxiv.org\/abs\/1803.09010"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/3392862"},{"key":"e_1_3_2_1_41_1","volume-title":"Splintering Urbanism: Networked Infrastructures, Technological Mobilities and the Urban Condition","author":"Graham Stephen","year":"2001","unstructured":"Stephen Graham and Simon Marvin. 2001. Splintering Urbanism: Networked Infrastructures, Technological Mobilities and the Urban Condition. Routledge, London."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372840"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3178876.3186138"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445929"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"crossref","unstructured":"Donna Haraway. 1988. Situated knowledges: The science question in feminism and the privilege of partial perspective. Feminist studies 14 3: 575\u2013599.","DOI":"10.2307\/3178066"},{"key":"e_1_3_2_1_46_1","volume-title":"Facebook is backing away from facial recognition. Meta isn't. Vox. Retrieved","author":"Heilweil Rebecca","year":"2022","unstructured":"Rebecca Heilweil. 2021. Facebook is backing away from facial recognition. Meta isn't. Vox. Retrieved January 9, 2022 from https:\/\/www.vox.com\/recode\/22761598\/facebook-facial-recognition-meta"},{"key":"e_1_3_2_1_47_1","volume-title":"The Guardian. Retrieved","author":"Hern Alex","year":"2020","unstructured":"Alex Hern. 2020. Ofqual's A-level algorithm: why did it fail to make the grade? The Guardian. Retrieved January 14, 2022 from https:\/\/www.theguardian.com\/education\/2020\/aug\/21\/ofqual-exams-algorithm-why-did-it-fail-make-grade-a-levels"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445918"},{"key":"e_1_3_2_1_49_1","volume-title":"Engineering play: A cultural history of children's software","author":"Ito Mizuko","unstructured":"Mizuko Ito. 2012. Engineering play: A cultural history of children's software. MIT Press."},{"key":"e_1_3_2_1_50_1","volume-title":"Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency., 375\u2013385","author":"Abigail","year":"2019","unstructured":"Abigail Z. Jacobs and Hanna Wallach. 2021. Measurement and Fairness. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency., 375\u2013385. Retrieved December 25, 2019 from http:\/\/arxiv.org\/abs\/1912.05511"},{"key":"e_1_3_2_1_51_1","unstructured":"Sheila Jasanoff (ed.). 2004. States of knowledge: the co-production of science and social order. Routledge London\u202f; New York."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445919"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"crossref","unstructured":"Michael Katell Meg Young Dharma Dailey Bernease Herman Vivian Guetler Aaron Tam Corinne Binz Daniella Raz and P M Krafft. 2020. Toward Situated Interventions for Algorithmic Equity: Lessons from the Field. 11.","DOI":"10.1145\/3351095.3372874"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1177\/2053951715621569"},{"key":"e_1_3_2_1_55_1","volume-title":"Software and everyday life","author":"Kitchin Rob","unstructured":"Rob Kitchin and Martin Dodge. 2014. Code\/space: Software and everyday life. Mit Press."},{"key":"e_1_3_2_1_56_1","volume-title":"Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks. Retrieved","author":"Koch Bernard","year":"2021","unstructured":"Bernard Koch, Emily Denton, Alex Hanna, and Jacob Gates Foster. 2021. Reduced, Reused and Recycled: The Life of a Dataset in Machine Learning Research. Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks. Retrieved December 13, 2021 from https:\/\/openreview.net\/forum?id=zNQBIBKJRkd"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287590"},{"key":"e_1_3_2_1_58_1","volume-title":"Standards and their Stories: How Quantifying, Classifying, and Formalizing Practices shape Everyday Life","author":"Lampland Martha","unstructured":"Martha Lampland and Susan Leigh Star. 2009. Standards and their Stories: How Quantifying, Classifying, and Formalizing Practices shape Everyday Life. Cornell University Press, Ithaca."},{"key":"e_1_3_2_1_59_1","volume-title":"Signal and Noise: Media, Infrastructure, and Urban Culture in Nigeria","author":"Larkin Brian","unstructured":"Brian Larkin. 2008. Signal and Noise: Media, Infrastructure, and Urban Culture in Nigeria. Duke University Press, Durham, NC."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/1180875.1180950"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1177\/2053951718756684"},{"key":"e_1_3_2_1_62_1","volume-title":"Proceedings of the ACM on Human-Computer Interaction 3, CSCW: 1\u201335","author":"Noothigattu Ritesh","year":"2019","unstructured":"Min Kyung Lee, Daniel Kusbit, Anson Kahng, Ji Tae Kim, Xinran Yuan, Allissa Chan, Daniel See, Ritesh Noothigattu, Siheon Lee, Alexandros Psomas, and others. 2019. WeBuildAI: Participatory framework for algorithmic governance. Proceedings of the ACM on Human-Computer Interaction 3, CSCW: 1\u201335."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.2307\/23043493"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.infoandorg.2013.02.002"},{"key":"e_1_3_2_1_65_1","volume-title":"Proceedings of the SIGCHI Conference on Human Factors in Computing Systems.","author":"Liao Q Vera","year":"2020","unstructured":"Q Vera Liao, Daniel Gruen, and Sarah Miller. 2020. Questioning the AI: Informing Design Practices for Explainable AI User Experiences. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems."},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376727"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10746-016-9385-5"},{"key":"e_1_3_2_1_68_1","unstructured":"Jonathan Magnusson. Improving Dark Pattern Literacy of End Users. 3."},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1353\/sor.2019.0022"},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445935"},{"key":"e_1_3_2_1_71_1","unstructured":"Sabelo Mhlambi. 2020. From Rationality to Relationality: Ubuntu as an Ethical and Human Rights Framework for Artificial Intelligence Governance. Retrieved from https:\/\/carrcenter.hks.harvard.edu\/publications\/rationality-relationality-ubuntu-ethical-and-human-rights-framework-artificial"},{"key":"e_1_3_2_1_72_1","volume-title":"The Oxford Handbook of Sociology and Digital Media, Deana A","author":"Milan Stefania","unstructured":"Stefania Milan and Emiliano Trer\u00e9. 2020. Big Data from the South(s): An Analytical Matrix to Investigate Data at the Margins. In The Oxford Handbook of Sociology and Digital Media, Deana A. Rohlinger and Sarah Sobieraj (eds.). Oxford University Press, Oxford."},{"key":"e_1_3_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287596"},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1146\/annurev-statistics-042720-125902"},{"key":"e_1_3_2_1_75_1","volume-title":"The Importance of AS-Level. Retrieved","author":"Moon Matthew","year":"2022","unstructured":"Matthew Moon. 2015. The Importance of AS-Level. Retrieved January 9, 2022 from https:\/\/www.cao.cam.ac.uk\/behind-the-headlines\/importance-of-aslevel"},{"key":"e_1_3_2_1_76_1","volume-title":"Ranjit Singh, Madeleine Clare Elish, and Jacob Metcalf.","author":"Moss Emanuel","year":"2021","unstructured":"Emanuel Moss, Elizabeth Anne Watkins, Ranjit Singh, Madeleine Clare Elish, and Jacob Metcalf. 2021. Assembling Accountability Through Algorithmic Impact Assessment. Data & Society Research Institute. Retrieved from http:\/\/datasociety.net\/library\/assembling-accountability\/"},{"key":"e_1_3_2_1_77_1","volume-title":"Proc. Conf. Fairness Accountability Transp.","author":"Narayanan Arvind","year":"2018","unstructured":"Arvind Narayanan. 2018. Translation tutorial: 21 fairness definitions and their politics. In Proc. Conf. Fairness Accountability Transp., New York, USA."},{"key":"e_1_3_2_1_78_1","volume-title":"Wired UK. Retrieved","author":"Nast Cond\u00e9","year":"2022","unstructured":"Cond\u00e9 Nast. Everything that went wrong with the botched A-Levels algorithm. Wired UK. Retrieved January 14, 2022 from https:\/\/www.wired.co.uk\/article\/alevel-exam-algorithm"},{"key":"e_1_3_2_1_79_1","volume-title":"Algorithms of Oppression: How Search Engines Reinforce Racism","author":"Noble Safiya Umoja","unstructured":"Safiya Umoja Noble. 2018. Algorithms of Oppression: How Search Engines Reinforce Racism. NYU Press, New York."},{"key":"e_1_3_2_1_80_1","doi-asserted-by":"publisher","unstructured":"Ziad Obermeyer Brian Powers Christine Vogeli and Sendhil Mullainathan. 2019. Dissecting racial bias in an algorithm used to manage the health of populations. Science (New York N.Y.) 366 6464: 447\u2013453. https:\/\/doi.org\/10.1126\/science.aax2342","DOI":"10.1126\/science.aax2342"},{"key":"e_1_3_2_1_81_1","volume-title":"The Office of Qualifications and Examinations Regulation (Ofqual) | About us. GOV.UK. Retrieved","year":"2022","unstructured":"Ofqual. The Office of Qualifications and Examinations Regulation (Ofqual) | About us. GOV.UK. Retrieved January 5, 2022 from https:\/\/www.gov.uk\/government\/organisations\/ofqual\/about"},{"key":"e_1_3_2_1_82_1","volume-title":"GOV.UK. Retrieved","year":"2021","unstructured":"Ofqual. Written statement from Chair of Ofqual to the Education Select Committee. GOV.UK. Retrieved October 28, 2021 from https:\/\/www.gov.uk\/government\/news\/written-statement-from-chair-of-ofqual-to-the-education-select-committee"},{"key":"e_1_3_2_1_83_1","doi-asserted-by":"publisher","DOI":"10.1093\/cje"},{"key":"e_1_3_2_1_84_1","volume-title":"The Verge. Retrieved","author":"Porter Jon","year":"2020","unstructured":"Jon Porter. 2020. UK ditches exam results generated by biased algorithm after student protests. The Verge. Retrieved January 14, 2022 from https:\/\/www.theverge.com\/2020\/8\/17\/21372045\/uk-a-level-results-algorithm-biased-coronavirus-covid-19-pandemic-university-applications"},{"key":"e_1_3_2_1_85_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372828"},{"key":"e_1_3_2_1_86_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372873"},{"key":"e_1_3_2_1_87_1","volume-title":"Engines of order: A mechanology of algorithmic techniques","author":"Rieder Bernhard","unstructured":"Bernhard Rieder. 2020. Engines of order: A mechanology of algorithmic techniques. Amsterdam University Press."},{"key":"e_1_3_2_1_88_1","doi-asserted-by":"crossref","unstructured":"Mary Beth Rosson and John M Carroll. 2009. Scenario based design. Human-computer interaction. boca raton FL: 145\u2013162.","DOI":"10.1201\/9781420088892.ch8"},{"key":"e_1_3_2_1_89_1","volume-title":"An Algorithmic Audit","author":"Sandvig Christian","unstructured":"Christian Sandvig, Kevin Hamilton, Karrie Karahalios, and Cedric Langbort. 2014. An Algorithmic Audit. In An Algorithmic Audit, Gangadharan (ed.). Open Technology Institute, available at: https:\/\/www. newamerica. org\/downloads\/OTI-Dataan-Discrimination-FINAL-small. pdf (accessed 3 July 2015), Data and Discrimination: Collected Essays. Retrieved October 16, 2015 from http:\/\/www.kevinhamilton.org\/share\/papers\/OTI-Data-an-Discrimination-FINAL-small.pdf"},{"key":"e_1_3_2_1_90_1","volume-title":"New Data Shows It Perpetuates Them. The Markup. Retrieved","author":"Sankin Aaron","year":"2021","unstructured":"Aaron Sankin, Dhruv Mehrotra for Gizmodo, Surya Mattu, and Annie Gilbertson. 2021. Crime Prediction Software Promised to Be Free of Biases. New Data Shows It Perpetuates Them. The Markup. Retrieved December 13, 2021 from https:\/\/themarkup.org\/prediction-bias\/2021\/12\/02\/crime-prediction-software-promised-to-be-free-of-biases-new-data-shows-it-perpetuates-them"},{"key":"e_1_3_2_1_91_1","doi-asserted-by":"publisher","DOI":"10.1177\/2053951717738104"},{"key":"e_1_3_2_1_92_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287598"},{"key":"e_1_3_2_1_93_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372827"},{"key":"e_1_3_2_1_94_1","doi-asserted-by":"publisher","DOI":"10.1145\/1094562.1094569"},{"key":"e_1_3_2_1_95_1","doi-asserted-by":"publisher","DOI":"10.1145\/3479577"},{"key":"e_1_3_2_1_96_1","doi-asserted-by":"publisher","DOI":"10.1177\/20539517211040759"},{"key":"e_1_3_2_1_97_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1618211113"},{"key":"e_1_3_2_1_98_1","doi-asserted-by":"crossref","unstructured":"AbdouMaliq Simone. 2004. People as infrastructure: Intersecting fragments in Johannesburg. Public culture 16 3: 407\u2013429.","DOI":"10.1215\/08992363-16-3-407"},{"key":"e_1_3_2_1_99_1","volume-title":"Lives of Data: Essays on Computational Culture in India","author":"Singh Ranjit","year":"2021","unstructured":"Ranjit Singh. 2020. Study the Imbrication: A Methodological Maxim to Follow the Multiple Lives of Data. In Lives of Data: Essays on Computational Culture in India, Sandeep Mertia (ed.). Institute of Network Cultures, Amsterdam, 51\u201359. Retrieved from https:\/\/networkcultures.org\/wp-content\/uploads\/2021\/02\/Lives-of-Data-.pdf"},{"key":"e_1_3_2_1_100_1","doi-asserted-by":"publisher","DOI":"10.1145\/3476056"},{"key":"e_1_3_2_1_101_1","volume-title":"Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems.","author":"Singh Ranjit","unstructured":"Ranjit Singh and Steven J. Jackson. 2017. From Margins to Seams: Imbrication, Inclusion, and Torque in the Aadhaar Identification Project. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems."},{"key":"e_1_3_2_1_102_1","volume-title":"A Sociology of Monsters: Essays on Power","author":"Star Susan Leigh","unstructured":"Susan Leigh Star. 1991. Power, Technology, and the Phenomenology of Conventions: On being Allergic to Onions. In A Sociology of Monsters: Essays on Power, Technology, and Domination, John Law (ed.). Routledge, London, 26\u201356."},{"key":"e_1_3_2_1_103_1","doi-asserted-by":"publisher","DOI":"10.1177\/00027649921955326"},{"key":"e_1_3_2_1_104_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.7.1.111"},{"key":"e_1_3_2_1_105_1","volume-title":"Technology's Storytellers: Reweaving the Human Fabric","author":"Staudenmaier John M","unstructured":"John M Staudenmaier. 1985. Technology's Storytellers: Reweaving the Human Fabric. The Society for the History of Technology and the MIT Press, Cambridge, MA."},{"key":"e_1_3_2_1_106_1","volume-title":"Human-machine reconfigurations: Plans and situated actions","author":"Suchman Lucy","unstructured":"Lucy Suchman and Lucy A Suchman. 2007. Human-machine reconfigurations: Plans and situated actions. Cambridge university press."},{"key":"e_1_3_2_1_107_1","first-page":"264","article-title":"Seamful Spaces","volume":"39","author":"Vertesi Janet","year":"2014","unstructured":"Janet Vertesi. 2014. Seamful Spaces: Heterogeneous Infrastructures in Interaction. Science, Technology & Human Values 39, 2: 264\u2013284. Retrieved from http:\/\/sth.sagepub.com\/content\/39\/2\/264.abstract","journal-title":"Heterogeneous Infrastructures in Interaction. Science, Technology & Human Values"},{"key":"e_1_3_2_1_108_1","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.3727562"},{"key":"e_1_3_2_1_109_1","doi-asserted-by":"publisher","DOI":"10.1145\/3461702.3462580"},{"key":"e_1_3_2_1_110_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3174230"},{"key":"e_1_3_2_1_111_1","doi-asserted-by":"publisher","DOI":"10.1177\/20539517211061122"},{"key":"e_1_3_2_1_112_1","volume-title":"All English medium schools must have government registration: Ministry. The Business Standard. Retrieved","year":"2022","unstructured":"2020. All English medium schools must have government registration: Ministry. The Business Standard. Retrieved January 9, 2022 from https:\/\/www.tbsnews.net\/bangladesh\/education\/all-english-medium-schools-must-have-government-registration-ministry-113293"},{"key":"e_1_3_2_1_113_1","volume-title":"Impact of Social Sciences. Retrieved","year":"2022","unstructured":"2020. \u201cF**k the algorithm\u201d?: What the world can learn from the UK's A-level grading fiasco. Impact of Social Sciences. Retrieved January 14, 2022 from https:\/\/blogs.lse.ac.uk\/impactofsocialsciences\/2020\/08\/26\/fk-the-algorithm-what-the-world-can-learn-from-the-uks-a-level-grading-fiasco\/"},{"key":"e_1_3_2_1_114_1","volume-title":"Facebook to Shut Down Use of Facial Recognition Technology. Bloomberg.com. Retrieved","year":"2022","unstructured":"2021. Facebook to Shut Down Use of Facial Recognition Technology. Bloomberg.com. Retrieved January 9, 2022 from https:\/\/www.bloomberg.com\/news\/articles\/2021-11-02\/facebook-to-shut-down-use-of-facial-recognition-technology"},{"key":"e_1_3_2_1_115_1","unstructured":"Where are Cambridge International AS & A Levels accepted and recognised? What can we help you with? Retrieved January 9 2022 from https:\/\/help.cambridgeinternational.org\/hc\/en-gb\/articles\/115004302785-Where-are-Cambridge-International-AS-A-Levels-accepted-and-recognised-"},{"key":"e_1_3_2_1_116_1","volume-title":"GOV.UK. Retrieved","author":"Guide","year":"2022","unstructured":"Guide to AS and A level results for England, 2019. GOV.UK. Retrieved January 14, 2022 from https:\/\/www.gov.uk\/government\/news\/guide-to-as-and-a-level-results-for-england-2019"},{"key":"e_1_3_2_1_117_1","volume-title":"AS and A levels: guide for schools and colleges. GOV.UK. Retrieved","author":"Es Regulating","year":"2022","unstructured":"Regulating GCSEs, AS and A levels: guide for schools and colleges. GOV.UK. Retrieved January 9, 2022 from https:\/\/www.gov.uk\/guidance\/regulating-gcses-as-and-a-levels-guide-for-schools-and-colleges"},{"key":"e_1_3_2_1_118_1","unstructured":"IGCSE | British Council. Retrieved January 9 2022 from https:\/\/www.britishcouncil.org.bd\/en\/exam\/igcse-school"},{"key":"e_1_3_2_1_119_1","unstructured":"Exam registration for all candidates | British Council. Retrieved January 9 2022 from https:\/\/www.britishcouncil.org.bd\/en\/exam\/igcse-school\/register\/private-candidates"},{"key":"e_1_3_2_1_120_1","unstructured":"Facebook reaches $550 million settlement in facial recognition lawsuit. NBC News. Retrieved January 9 2022 from https:\/\/www.nbcnews.com\/tech\/tech-news\/facebook-reaches-550-million-settlement-facial-recognition-lawsuit-n1126191"}],"event":{"name":"FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency","location":"Seoul Republic of Korea","acronym":"FAccT '22","sponsor":["ACM Association for Computing Machinery"]},"container-title":["2022 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533186","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3533186","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3533186","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:02:09Z","timestamp":1750186929000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533186"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,20]]},"references-count":120,"alternative-id":["10.1145\/3531146.3533186","10.1145\/3531146"],"URL":"https:\/\/doi.org\/10.1145\/3531146.3533186","relation":{},"subject":[],"published":{"date-parts":[[2022,6,20]]},"assertion":[{"value":"2022-06-20","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}