{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T20:14:46Z","timestamp":1776111286593,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":117,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T00:00:00Z","timestamp":1655683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,21]]},"DOI":"10.1145\/3531146.3533237","type":"proceedings-article","created":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T14:27:10Z","timestamp":1655735230000},"page":"1917-1928","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":35,"title":["How Platform-User Power Relations Shape Algorithmic Accountability"],"prefix":"10.1145","author":[{"given":"Divya","family":"Ramesh","sequence":"first","affiliation":[{"name":"Computer Science and Engineering, University of Michigan, Ann Arbor, USA"}]},{"given":"Vaishnav","family":"Kameswaran","sequence":"additional","affiliation":[{"name":"School of Information, University of Michigan, Ann Arbor, USA"}]},{"given":"Ding","family":"Wang","sequence":"additional","affiliation":[{"name":"Google Research, India"}]},{"given":"Nithya","family":"Sambasivan","sequence":"additional","affiliation":[{"name":"Unaffiliated, USA"}]}],"member":"320","published-online":{"date-parts":[[2022,6,20]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"2021. Non-Banking Financial Company. https:\/\/www.rbi.org.in\/Scripts\/FAQView.aspx?Id=92"},{"key":"e_1_3_2_1_2_1","unstructured":"2022. Dhani - India\u2019s Trusted Site for Finance Healthcare and Online Medicines. https:\/\/www.dhani.com\/"},{"key":"e_1_3_2_1_3_1","unstructured":"2022. Five ways that AI augments FinTech. https:\/\/indiaai.gov.in\/article\/five-ways-that-ai-augments-fintech"},{"key":"e_1_3_2_1_4_1","unstructured":"2022. Get line of credit up to Rs. 5 Lakhs - MoneyTap. https:\/\/www.moneytap.com"},{"key":"e_1_3_2_1_5_1","unstructured":"2022. Kissht. https:\/\/kissht.com"},{"key":"e_1_3_2_1_6_1","unstructured":"2022. KreditBee. https:\/\/www.kreditbee.in\/"},{"key":"e_1_3_2_1_7_1","unstructured":"2022. SmartCoin - Get Instant Credit. https:\/\/smartcoin.co.in\/"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3174156"},{"key":"e_1_3_2_1_9_1","volume-title":"AI Now Institute, and Open Government Partnership","author":"Ada Lovelace Institute","year":"2021","unstructured":"Ada Lovelace Institute, AI Now Institute, and Open Government Partnership. 2021. Algorithmic Accountability for the Public Sector. (2021)."},{"key":"e_1_3_2_1_10_1","volume-title":"Peeking inside the black-box: a survey on explainable artificial intelligence (XAI)","author":"Adadi Amina","year":"2018","unstructured":"Amina Adadi and Mohammed Berrada. 2018. Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE access 6(2018), 52138\u201352160."},{"key":"e_1_3_2_1_11_1","unstructured":"Sray Agarwal. 2020. AI powered FinTech: The drivers of Digital India. https:\/\/indiaai.gov.in\/article\/ai-powered-fintech-the-drivers-of-digital-india"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3025453.3025961"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/2556288.2557376"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3409557"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300233"},{"key":"e_1_3_2_1_16_1","unstructured":"Amsterdam. 2020. Algorithmic Register Amsterdam. https:\/\/algoritmeregister.amsterdam.nl\/en\/ai-register\/"},{"key":"e_1_3_2_1_17_1","volume-title":"Seeing without knowing: Limitations of the transparency ideal and its application to algorithmic accountability. new media & society 20, 3","author":"Ananny Mike","year":"2018","unstructured":"Mike Ananny and Kate Crawford. 2018. Seeing without knowing: Limitations of the transparency ideal and its application to algorithmic accountability. new media & society 20, 3 (2018), 973\u2013989."},{"key":"e_1_3_2_1_18_1","volume-title":"suicide and the dodgy loan apps plaguing Google\u2019s Play Store. Wired UK","author":"Bansal Varsha","year":"2021","unstructured":"Varsha Bansal. 2021. Shame, suicide and the dodgy loan apps plaguing Google\u2019s Play Store. Wired UK (2021). https:\/\/www.wired.co.uk\/article\/google-loan-apps-india-deaths"},{"key":"e_1_3_2_1_19_1","volume-title":"Showing Limits of Algorithms. Wall Street Journal (jul","author":"Barr Alistair","year":"2015","unstructured":"Alistair Barr. 2015. Google Mistakenly Tags Black People as \u2018Gorillas,\u2019 Showing Limits of Algorithms. Wall Street Journal (jul 2015). https:\/\/www.wsj.com\/articles\/BL-DGB-42522"},{"key":"e_1_3_2_1_20_1","volume-title":"Race after technology: Abolitionist tools for the new jim code. Social Forces","author":"Benjamin Ruha","year":"2019","unstructured":"Ruha Benjamin. 2019. Race after technology: Abolitionist tools for the new jim code. Social Forces (2019)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3173951"},{"key":"e_1_3_2_1_22_1","volume-title":"Man is to computer programmer as woman is to homemaker? debiasing word embeddings. Advances in neural information processing systems 29","author":"Bolukbasi Tolga","year":"2016","unstructured":"Tolga Bolukbasi, Kai-Wei Chang, James\u00a0Y Zou, Venkatesh Saligrama, and Adam\u00a0T Kalai. 2016. Man is to computer programmer as woman is to homemaker? debiasing word embeddings. Advances in neural information processing systems 29 (2016), 4349\u20134357."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1080\/1369118X.2012.678878"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"crossref","unstructured":"Virginia Braun and Victoria Clarke. 2012. Thematic analysis.(2012).","DOI":"10.1037\/13620-004"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300271"},{"key":"e_1_3_2_1_26_1","volume-title":"Conference on fairness, accountability and transparency. PMLR, 77\u201391","author":"Buolamwini Joy","year":"2018","unstructured":"Joy Buolamwini and Timnit Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency. PMLR, 77\u201391."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1177\/2053951715622512"},{"key":"e_1_3_2_1_28_1","unstructured":"CAC. 2021. Internet Information Service Algorithm Recommendation Management Regulations. http:\/\/www.cac.gov.cn\/2021-08\/27\/c_1631652502874117.htm."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359206"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.3390\/electronics8080832"},{"key":"e_1_3_2_1_32_1","unstructured":"C Chausson. 2016. France opens the source code of tax and benefits calculators to increase transparency."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"crossref","unstructured":"Sasha Costanza-Chock. 2020. Design justice: Community-led practices to build the worlds we need.","DOI":"10.7551\/mitpress\/12255.001.0001"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403363"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1080\/21670811.2014.976411"},{"key":"e_1_3_2_1_36_1","volume-title":"Transparent data mining for Big and Small Data","author":"Diakopoulos Nicholas","unstructured":"Nicholas Diakopoulos. 2017. Enabling accountability of algorithmic media: transparency as a constructive and critical lens. In Transparent data mining for Big and Small Data. Springer, 25\u201343."},{"key":"e_1_3_2_1_37_1","volume-title":"Algorithmic transparency in the news media. Digital journalism 5, 7","author":"Diakopoulos Nicholas","year":"2017","unstructured":"Nicholas Diakopoulos and Michael Koliska. 2017. Algorithmic transparency in the news media. Digital journalism 5, 7 (2017), 809\u2013828."},{"key":"e_1_3_2_1_38_1","volume-title":"Data feminism","author":"D\u2019ignazio Catherine","unstructured":"Catherine D\u2019ignazio and Lauren\u00a0F Klein. 2020. Data feminism. MIT press."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/1858171.1858173"},{"key":"e_1_3_2_1_40_1","volume-title":"The accuracy, fairness, and limits of predicting recidivism. Science advances","author":"Dressel Julia","year":"2018","unstructured":"Julia Dressel and Hany Farid. 2018. The accuracy, fairness, and limits of predicting recidivism. Science advances (2018)."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3025453.3025728"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445188"},{"key":"e_1_3_2_1_43_1","unstructured":"Upol Ehsan Samir Passi Q\u00a0Vera Liao Larry Chan I Lee Michael Muller Mark\u00a0O Riedl 2021. The Who in Explainable AI: How AI Background Shapes Perceptions of AI Explanations. arXiv preprint arXiv:2107.13509(2021)."},{"key":"e_1_3_2_1_44_1","unstructured":"MC Elish and EA Watkins. 2020. Repairing innovation: A study of integrating AI in clinical care. Unpublished Manuscript(2020)."},{"key":"e_1_3_2_1_45_1","volume-title":"Could machine learning help companies select better board directors?Harvard Business Review 1, 5","author":"Erel Isil","year":"2018","unstructured":"Isil Erel, Lea\u00a0H Stern, Chenhao Tan, and Michael\u00a0S Weisbach. 2018. Could machine learning help companies select better board directors?Harvard Business Review 1, 5 (2018)."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3392874"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/2702123.2702556"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300724"},{"key":"e_1_3_2_1_49_1","unstructured":"ET Goverment. 2021. Odisha launches AI based online life certificate system for pensioners."},{"key":"e_1_3_2_1_50_1","volume-title":"Automating inequality: How high-tech tools profile, police, and punish the poor","author":"Eubanks Virginia","unstructured":"Virginia Eubanks. 2018. Automating inequality: How high-tech tools profile, police, and punish the poor. St. Martin\u2019s Press."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"crossref","unstructured":"Simson Garfinkel Jeanna Matthews Stuart\u00a0S Shapiro and Jonathan\u00a0M Smith. 2017. Toward algorithmic transparency and accountability.","DOI":"10.1145\/3125780"},{"key":"e_1_3_2_1_52_1","volume-title":"Automated underwriting in mortgage lending: Good news for the underserved?Housing Policy Debate 13, 2","author":"Gates Susan\u00a0Wharton","year":"2002","unstructured":"Susan\u00a0Wharton Gates, Vanessa\u00a0Gail Perry, and Peter\u00a0M Zorn. 2002. Automated underwriting in mortgage lending: Good news for the underserved?Housing Policy Debate 13, 2 (2002), 369\u2013391."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3458723"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.7551\/mitpress\/9780262525374.001.0001"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/DSAA.2018.00018"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1001\/jama.2016.17216"},{"key":"e_1_3_2_1_59_1","volume-title":"L van\u00a0de Fliert, and P Rautio","author":"Haataja M","year":"2020","unstructured":"M Haataja, L van\u00a0de Fliert, and P Rautio. 2020. Public AI Registers: Realising AI transparency and civic participation in government use of AI. (2020)."},{"key":"e_1_3_2_1_60_1","unstructured":"Alexa Hagerty and Igor Rubinov. 2019. Global AI ethics: a review of the social impacts and ethical implications of artificial intelligence. arXiv preprint arXiv:1907.07892(2019)."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2945386"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1145\/3461702.3462564"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1080\/1369118X.2019.1573912"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445918"},{"key":"e_1_3_2_1_65_1","unstructured":"ICF IIPS. 2017. India National Family Health Survey NFHS-4 2015\u201316. Mumbai: IIPS and ICF(2017)."},{"key":"e_1_3_2_1_66_1","unstructured":"Google India. 2022. Helping users stay safe online. https:\/\/forindia.withgoogle.com\/intl\/en\/trust-and-safety\/"},{"key":"e_1_3_2_1_67_1","unstructured":"Internet Freedom Foundation. [n.d.]. Project Panoptic: Facial Recognition Systems in India."},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.1145\/2470654.2470742"},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1177\/20539517211020332"},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445931"},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1145\/3278721.3278738"},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359199"},{"key":"e_1_3_2_1_73_1","volume-title":"User Attitudes and Sources of AI Authority in India. In CHI Conference on Human Factors in Computing Systems (CHI \u201922)","author":"Kapania Shivani","year":"2022","unstructured":"Shivani Kapania, Oliver Siy, Gabe Clapper, Azhagu SP, and Nithya Sambasivan. 2022. \u201dBecause AI is 100% right and safe\u201d: User Attitudes and Sources of AI Authority in India. In CHI Conference on Human Factors in Computing Systems (CHI \u201922)."},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372874"},{"key":"e_1_3_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376219"},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1080\/1369118X.2018.1477967"},{"key":"e_1_3_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290607.3310433"},{"key":"e_1_3_2_1_78_1","doi-asserted-by":"publisher","DOI":"10.1145\/2493432.2493474"},{"key":"e_1_3_2_1_79_1","volume-title":"Doing interviews","author":"Kvale Steinar","unstructured":"Steinar Kvale. 2008. Doing interviews. Sage."},{"key":"e_1_3_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1145\/2702123.2702548"},{"key":"e_1_3_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445570"},{"key":"e_1_3_2_1_83_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376590"},{"key":"e_1_3_2_1_84_1","volume-title":"Label on Video of Black Men. The New York Times (sep","author":"Mac Ryan","year":"2021","unstructured":"Ryan Mac. 2021. Facebook Apologizes After A.I. Puts \u2018Primates\u2019 Label on Video of Black Men. The New York Times (sep 2021). https:\/\/www.nytimes.com\/2021\/09\/03\/technology\/facebook-ai-race-primates.html"},{"key":"e_1_3_2_1_86_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372865"},{"key":"e_1_3_2_1_87_1","doi-asserted-by":"publisher","DOI":"10.1145\/3457607"},{"key":"e_1_3_2_1_88_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445935"},{"key":"e_1_3_2_1_89_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287596"},{"key":"e_1_3_2_1_90_1","volume-title":"Assembling Accountability: Algorithmic Impact Assessment for the Public Interest. Available at SSRN 3877437(2021).","author":"Moss Emanuel","year":"2021","unstructured":"Emanuel Moss, Elizabeth\u00a0Anne Watkins, Ranjit Singh, Madeleine\u00a0Clare Elish, and Jacob Metcalf. 2021. Assembling Accountability: Algorithmic Impact Assessment for the Public Interest. Available at SSRN 3877437(2021)."},{"key":"e_1_3_2_1_91_1","unstructured":"NITI Aayog. 2021. Responsible AI #AIFORALL. (2021)."},{"key":"e_1_3_2_1_92_1","volume-title":"Algorithms of oppression","author":"Noble Safiya\u00a0Umoja","unstructured":"Safiya\u00a0Umoja Noble. 2018. Algorithms of oppression. New York University Press."},{"key":"e_1_3_2_1_93_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445420"},{"key":"e_1_3_2_1_94_1","unstructured":"Cathy O\u2019neil. 2016. Weapons of math destruction: How big data increases inequality and threatens democracy. Crown."},{"key":"e_1_3_2_1_95_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10606-017-9289-6"},{"key":"e_1_3_2_1_96_1","unstructured":"Google PAIR. 2019. People + AI Guidebook. https:\/\/design.google\/ai-guidebook"},{"key":"e_1_3_2_1_97_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3173803"},{"key":"e_1_3_2_1_98_1","volume-title":"The black box society","author":"Pasquale Frank","unstructured":"Frank Pasquale. 2015. The black box society. Harvard University Press."},{"key":"e_1_3_2_1_99_1","unstructured":"S P\u00e9nicaud. 2021. Building Public Algorithm Registers: Lessons Learned from the French Approach. (2021)."},{"key":"e_1_3_2_1_100_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3173677"},{"key":"e_1_3_2_1_101_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306618.3314244"},{"key":"e_1_3_2_1_102_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372873"},{"key":"e_1_3_2_1_103_1","doi-asserted-by":"publisher","DOI":"10.1145\/3449081"},{"key":"e_1_3_2_1_104_1","doi-asserted-by":"publisher","DOI":"10.1525\/gp.2021.26132"},{"key":"e_1_3_2_1_105_1","unstructured":"Marco\u00a0Tulio Ribeiro Sameer Singh and Carlos Guestrin. 2016. Model-agnostic interpretability of machine learning. arXiv preprint arXiv:1606.05386(2016)."},{"key":"e_1_3_2_1_106_1","unstructured":"Rashida Richardson. 2021. Defining and Demystifying Automated Decision Systems. Maryland Law Review Forthcoming(2021)."},{"key":"e_1_3_2_1_107_1","volume-title":"When One Affects Many: The Case For Collective Consent","author":"Ruhaak A","year":"2021","unstructured":"A Ruhaak. 2021. When One Affects Many: The Case For Collective Consent. Mozilla Foundation 20(2021)."},{"key":"e_1_3_2_1_108_1","volume-title":"The coding manual for qualitative researchers","author":"Salda\u00f1a Johnny","unstructured":"Johnny Salda\u00f1a. 2015. The coding manual for qualitative researchers. Sage."},{"key":"e_1_3_2_1_109_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445896"},{"key":"e_1_3_2_1_110_1","doi-asserted-by":"publisher","DOI":"10.1145\/1753326.1753718"},{"key":"e_1_3_2_1_111_1","volume-title":"Data Cascades in High-Stakes AI. In proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1\u201315","author":"Sambasivan Nithya","year":"2021","unstructured":"Nithya Sambasivan, Shivani Kapania, Hannah Highfill, Diana Akrong, Praveen Paritosh, and Lora\u00a0M Aroyo. 2021. \u201cEveryone wants to do the model work, not the data work\u201d: Data Cascades in High-Stakes AI. In proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1\u201315."},{"key":"e_1_3_2_1_112_1","unstructured":"Christian Sandvig Kevin Hamilton Karrie Karahalios and Cedric Langbort. [n.d.]. Auditing algorithms: Research methods for detecting discrimination on internet platforms. ([n. d.])."},{"key":"e_1_3_2_1_113_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376624"},{"key":"e_1_3_2_1_114_1","volume-title":"Plans and situated actions: The problem of human-machine communication","author":"Suchman A","unstructured":"Lucy\u00a0A Suchman. 1987. Plans and situated actions: The problem of human-machine communication. Cambridge university press."},{"key":"e_1_3_2_1_116_1","volume-title":"Australian Uber drivers say the company is manipulating their ratings to boost its fees. Business Insider Australia 20","author":"Tucker Harry","year":"2016","unstructured":"Harry Tucker. 2016. Australian Uber drivers say the company is manipulating their ratings to boost its fees. Business Insider Australia 20 (2016)."},{"key":"e_1_3_2_1_117_1","unstructured":"James Vincent. 2016. Twitter taught {Microsoft}\u2019s friendly {AI} chatbot to be a racist asshole in less than a day. https:\/\/www.theverge.com\/2016\/3\/24\/11297050\/tay-microsoft-chatbot-racist"},{"key":"e_1_3_2_1_118_1","doi-asserted-by":"publisher","DOI":"10.1145\/2207676.2208567"},{"key":"e_1_3_2_1_119_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300831"},{"key":"e_1_3_2_1_120_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376813"},{"key":"e_1_3_2_1_121_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372833"},{"key":"e_1_3_2_1_122_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3174230"},{"key":"e_1_3_2_1_123_1","doi-asserted-by":"crossref","unstructured":"Kyra Yee Uthaipon Tantipongpipat and Shubhanshu Mishra. 2021. Image Cropping on Twitter: Fairness Metrics their Limitations and the Importance of Representation Design and Agency. arXiv preprint arXiv:2105.08667(2021).","DOI":"10.1145\/3479594"}],"event":{"name":"FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency","location":"Seoul Republic of Korea","acronym":"FAccT '22","sponsor":["ACM Association for Computing Machinery"]},"container-title":["2022 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533237","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3533237","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:31:30Z","timestamp":1750188690000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533237"}},"subtitle":["A Case Study of Instant Loan Platforms and Financially Stressed Users in India"],"short-title":[],"issued":{"date-parts":[[2022,6,20]]},"references-count":117,"alternative-id":["10.1145\/3531146.3533237","10.1145\/3531146"],"URL":"https:\/\/doi.org\/10.1145\/3531146.3533237","relation":{},"subject":[],"published":{"date-parts":[[2022,6,20]]},"assertion":[{"value":"2022-06-20","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}