{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,26]],"date-time":"2026-04-26T09:21:33Z","timestamp":1777195293857,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":69,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T00:00:00Z","timestamp":1686528000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,6,12]]},"DOI":"10.1145\/3593013.3594078","type":"proceedings-article","created":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T14:40:46Z","timestamp":1686580846000},"page":"1246-1266","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":31,"title":["\u201cI\u2019m fully who I am\u201d: Towards Centering Transgender and Non-Binary Voices to Measure Biases in Open Language Generation"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0531-7520","authenticated-orcid":false,"given":"Anaelia","family":"Ovalle","sequence":"first","affiliation":[{"name":"Computer Science, University of California, Los Angeles, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2455-2160","authenticated-orcid":false,"given":"Palash","family":"Goyal","sequence":"additional","affiliation":[{"name":"Alexa AI-NU, Amazon, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5396-9187","authenticated-orcid":false,"given":"Jwala","family":"Dhamala","sequence":"additional","affiliation":[{"name":"Alexa AI-NU, Amazon, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-1147-4332","authenticated-orcid":false,"given":"Zachary","family":"Jaggers","sequence":"additional","affiliation":[{"name":"Amazon Global Diversity, Equity, &amp; Inclusion, Amazon, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5365-0072","authenticated-orcid":false,"given":"Kai-Wei","family":"Chang","sequence":"additional","affiliation":[{"name":"Alexa AI-NU, Amazon, USA and Department of Computer Science, University of California, Los Angeles, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4215-0886","authenticated-orcid":false,"given":"Aram","family":"Galstyan","sequence":"additional","affiliation":[{"name":"Alexa AI-NU, Amazon, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9353-7509","authenticated-orcid":false,"given":"Richard","family":"Zemel","sequence":"additional","affiliation":[{"name":"Alexa AI-NU, Amazon, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9277-3718","authenticated-orcid":false,"given":"Rahul","family":"Gupta","sequence":"additional","affiliation":[{"name":"Alexa AI-NU, Amazon, USA"}]}],"member":"320","published-online":{"date-parts":[[2023,6,12]]},"reference":[{"key":"e_1_3_2_2_1_1","unstructured":"Social Security Administration. 2022. Popular Baby Names \u2014 ssa.gov. https:\/\/www.ssa.gov\/oact\/babynames\/index.html. [Accessed 05-Feb-2023]."},{"key":"e_1_3_2_2_2_1","unstructured":"AllenNLP. (n.d.). AllenNLP Demo \u2014 demo.allennlp.org. https:\/\/demo.allennlp.org\/coreference-resolution\/. [Accessed 26-Jan-2023]."},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.5172\/mra.2013.7.2.160"},{"key":"e_1_3_2_2_4_1","volume-title":"RedditBias: A real-world resource for bias evaluation and debiasing of conversational language models. arXiv preprint arXiv:2106.03521","author":"Barikeri Soumya","year":"2021","unstructured":"Soumya Barikeri, Anne Lauscher, Ivan Vuli\u0107, and Goran Glava\u0161. 2021. RedditBias: A real-world resource for bias evaluation and debiasing of conversational language models. arXiv preprint arXiv:2106.03521 (2021)."},{"key":"e_1_3_2_2_5_1","volume-title":"Fairness and machine learning: limitations and opportunities.[S. l.]: fairmlbook. org","author":"Solon BAROCAS","year":"2019","unstructured":"Solon BAROCAS, Moritz HARDT, and Arvind NARAYANAN. 2022. Fairness and machine learning: limitations and opportunities.[S. l.]: fairmlbook. org, 2019."},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1609\/icwsm.v14i1.7347"},{"key":"e_1_3_2_2_7_1","volume-title":"Singular they and the syntactic representation of gender in English. Glossa: a journal of general linguistics 2, 1","author":"Bjorkman Bronwyn M","year":"2017","unstructured":"Bronwyn M Bjorkman. 2017. Singular they and the syntactic representation of gender in English. Glossa: a journal of general linguistics 2, 1 (2017)."},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","unstructured":"Sid Black Leo Gao Phil Wang Connor Leahy and Stella Biderman. 2021. GPT-Neo: Large Scale Autoregressive Language Modeling with Mesh-Tensorflow. https:\/\/doi.org\/10.5281\/zenodo.5297715 If you use this software please cite it using these metadata..","DOI":"10.5281\/zenodo.5297715"},{"key":"e_1_3_2_2_9_1","volume-title":"Autumn Hamilton, and Eli Coleman.","author":"Bockting Walter O","year":"2013","unstructured":"Walter O Bockting, Michael H Miner, Rebecca E Swinburne Romine, Autumn Hamilton, and Eli Coleman. 2013. Stigma, mental health, and resilience in an online sample of the US transgender population. American journal of public health 103, 5 (2013), 943\u2013951."},{"key":"e_1_3_2_2_10_1","volume-title":"Man is to computer programmer as woman is to homemaker? debiasing word embeddings. Advances in neural information processing systems 29","author":"Bolukbasi Tolga","year":"2016","unstructured":"Tolga Bolukbasi, Kai-Wei Chang, James Y Zou, Venkatesh Saligrama, and Adam T Kalai. 2016. Man is to computer programmer as woman is to homemaker? debiasing word embeddings. Advances in neural information processing systems 29 (2016)."},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3404983.3405510"},{"key":"e_1_3_2_2_12_1","volume-title":"Semantics derived automatically from language corpora contain human-like biases. Science 356, 6334","author":"Caliskan Aylin","year":"2017","unstructured":"Aylin Caliskan, Joanna J Bryson, and Arvind Narayanan. 2017. Semantics derived automatically from language corpora contain human-like biases. Science 356, 6334 (2017), 183\u2013186."},{"key":"e_1_3_2_2_13_1","unstructured":"Cambridge. (n.d.). Determiners used as pronouns. https:\/\/dictionary.cambridge.org\/us\/grammar\/british-grammar\/determiners-used-as-pronouns."},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1162\/coli_a_00413"},{"key":"e_1_3_2_2_15_1","unstructured":"Gender Census. (n.d.). Gender Census 2022: Worldwide Report. https:\/\/www.gendercensus.com\/results\/2022-worldwide\/#pronouns. [Accessed 25-Jan-2023]."},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1300\/J082v51n03_04"},{"key":"e_1_3_2_2_17_1","unstructured":"St. Louis Community College. (n.d.). Pronoun and antecedent agreement. https:\/\/stlcc.edu\/student-support\/academic-success-and-tutoring\/writing-center\/writing-resources\/pronoun-antecedent-agreement.aspx"},{"key":"e_1_3_2_2_18_1","unstructured":"Kirby Conrod. 2019. Pronouns raising and emerging. Ph. D. Dissertation."},{"key":"e_1_3_2_2_19_1","volume-title":"Detecting Harmful Online Conversational Content towards LGBTQIA+ Individuals. arXiv preprint arXiv:2207.10032","author":"Dacon Jamell","year":"2022","unstructured":"Jamell Dacon, Harry Shomer, Shaylynn Crum-Dacon, and Jiliang Tang. 2022. Detecting Harmful Online Conversational Content towards LGBTQIA+ Individuals. arXiv preprint arXiv:2207.10032 (2022)."},{"key":"e_1_3_2_2_20_1","volume-title":"Why Can GPT Learn In-Context? Language Models Secretly Perform Gradient Descent as Meta Optimizers. arXiv preprint arXiv:2212.10559","author":"Dai Damai","year":"2022","unstructured":"Damai Dai, Yutao Sun, Li Dong, Yaru Hao, Zhifang Sui, and Furu Wei. 2022. Why Can GPT Learn In-Context? Language Models Secretly Perform Gradient Descent as Meta Optimizers. arXiv preprint arXiv:2212.10559 (2022)."},{"key":"e_1_3_2_2_21_1","volume-title":"Harms of gender exclusivity and challenges in non-binary representation in language technologies. arXiv preprint arXiv:2108.12084","author":"Dev Sunipa","year":"2021","unstructured":"Sunipa Dev, Masoud Monajatipoor, Anaelia Ovalle, Arjun Subramonian, Jeff M Phillips, and Kai-Wei Chang. 2021. Harms of gender exclusivity and challenges in non-binary representation in language technologies. arXiv preprint arXiv:2108.12084 (2021)."},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.findings-aacl.24"},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445924"},{"key":"e_1_3_2_2_24_1","volume-title":"A brief history of singular \u2018they","author":"Dictionary Oxford English","year":"2023","unstructured":"Oxford English Dictionary. (n.d.). A brief history of singular \u2018they\u2019 | Oxford English Dictionary \u2014 public.oed.com. https:\/\/public.oed.com\/blog\/a-brief-history-of-singular-they\/. [Accessed 25-Jan-2023]."},{"key":"e_1_3_2_2_25_1","volume-title":"Queens are powerful too: Mitigating gender bias in dialogue generation. arXiv preprint arXiv:1911.03842","author":"Dinan Emily","year":"2019","unstructured":"Emily Dinan, Angela Fan, Adina Williams, Jack Urbanek, Douwe Kiela, and Jason Weston. 2019. Queens are powerful too: Mitigating gender bias in dialogue generation. arXiv preprint arXiv:1911.03842 (2019)."},{"key":"e_1_3_2_2_26_1","volume-title":"A Survey for In-context Learning. arXiv preprint arXiv:2301.00234","author":"Dong Qingxiu","year":"2022","unstructured":"Qingxiu Dong, Lei Li, Damai Dai, Ce Zheng, Zhiyong Wu, Baobao Chang, Xu Sun, Jingjing Xu, and Zhifang Sui. 2022. A Survey for In-context Learning. arXiv preprint arXiv:2301.00234 (2022)."},{"key":"e_1_3_2_2_27_1","volume-title":"Eugene Jang, and Jonathan","author":"Felkner Virginia K","year":"2022","unstructured":"Virginia K Felkner, Ho-Chun Herbert Chang, Eugene Jang, and Jonathan May. 2022. Towards WinoQueer: Developing a Benchmark for Anti-Queer Bias in Large Language Models. arXiv preprint arXiv:2206.11484 (2022)."},{"key":"e_1_3_2_2_28_1","unstructured":"A Flowers. 2015. The most common unisex names in America: Is yours one of them? FiveThirtyEight."},{"key":"e_1_3_2_2_29_1","volume-title":"The pile: An 800gb dataset of diverse text for language modeling. arXiv preprint arXiv:2101.00027","author":"Gao Leo","year":"2020","unstructured":"Leo Gao, Stella Biderman, Sid Black, Laurence Golding, Travis Hoppe, Charles Foster, Jason Phang, Horace He, Anish Thite, Noa Nabeshima, 2020. The pile: An 800gb dataset of diverse text for language modeling. arXiv preprint arXiv:2101.00027 (2020)."},{"key":"e_1_3_2_2_30_1","unstructured":"Robin Hewings. (n.d.). Marginalization and Loneliness Among Sexual Minorities: How Are They Linked? - Campaign to End Loneliness \u2014 campaigntoendloneliness.org. https:\/\/www.campaigntoendloneliness.org\/marginalization-and-loneliness-among-sexual-minorities-how-are-they-linked. [Accessed 25-Jan-2023]."},{"key":"e_1_3_2_2_31_1","volume-title":"Deceiving google\u2019s perspective api built for detecting toxic comments. arXiv preprint arXiv:1702.08138","author":"Hosseini Hossein","year":"2017","unstructured":"Hossein Hosseini, Sreeram Kannan, Baosen Zhang, and Radha Poovendran. 2017. Deceiving google\u2019s perspective api built for detecting toxic comments. arXiv preprint arXiv:1702.08138 (2017)."},{"key":"e_1_3_2_2_32_1","unstructured":"HuggingFace. (n.d.). Neural Coreference. https:\/\/huggingface.co\/coref\/. [Accessed 26-Jan-2023]."},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1080\/15228835.2015.1105768"},{"key":"e_1_3_2_2_34_1","unstructured":"Sandy James Jody Herman Susan Rankin Mara Keisling Lisa Mottet and Ma\u2019ayan Anafi. 2016. The report of the 2015 US transgender survey. (2016)."},{"key":"e_1_3_2_2_35_1","volume-title":"Likert scale: Explored and explained. British journal of applied science & technology 7, 4","author":"Joshi Ankur","year":"2015","unstructured":"Ankur Joshi, Saket Kale, Satish Chandel, and D Kumar Pal. 2015. Likert scale: Explored and explained. British journal of applied science & technology 7, 4 (2015), 396."},{"key":"e_1_3_2_2_36_1","volume-title":"Welcome to the Modern World of Pronouns: Identity-Inclusive Natural Language Processing beyond Gender. arXiv preprint arXiv:2202.11923","author":"Lauscher Anne","year":"2022","unstructured":"Anne Lauscher, Archie Crowley, and Dirk Hovy. 2022. Welcome to the Modern World of Pronouns: Identity-Inclusive Natural Language Processing beyond Gender. arXiv preprint arXiv:2202.11923 (2022)."},{"key":"e_1_3_2_2_37_1","volume-title":"Mitigating gender bias for neural dialogue generation with adversarial learning. arXiv preprint arXiv:2009.13028","author":"Liu Haochen","year":"2020","unstructured":"Haochen Liu, Wentao Wang, Yiqi Wang, Hui Liu, Zitao Liu, and Jiliang Tang. 2020. Mitigating gender bias for neural dialogue generation with adversarial learning. arXiv preprint arXiv:2009.13028 (2020)."},{"key":"e_1_3_2_2_38_1","volume-title":"What Makes Good In-Context Examples for GPT-3 ?arXiv preprint arXiv:2101.06804","author":"Liu Jiachang","year":"2021","unstructured":"Jiachang Liu, Dinghan Shen, Yizhe Zhang, Bill Dolan, Lawrence Carin, and Weizhu Chen. 2021. What Makes Good In-Context Examples for GPT-3 ?arXiv preprint arXiv:2101.06804 (2021)."},{"key":"e_1_3_2_2_39_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.nuse-1.5"},{"key":"e_1_3_2_2_41_1","volume-title":"A minority stress perspective on transgender individuals","author":"McLemore Kevin A","year":"2018","unstructured":"Kevin A McLemore. 2018. A minority stress perspective on transgender individuals\u2019 experiences with misgendering.Stigma and Health 3, 1 (2018), 53."},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.191"},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.ltedi-1.4"},{"key":"e_1_3_2_2_44_1","unstructured":"OpenAI. 2023. ChatGPT: Optimizing language models for dialogue. https:\/\/openai.com\/blog\/chatgpt\/"},{"key":"e_1_3_2_2_45_1","unstructured":"Pearson. (n.d.). Gender Policing and Gender Accountability. https:\/\/revelpreview.pearson.com\/epubs\/pearson_kimmel_soc\/OPS\/xhtml\/ch09_pg0013.xhtml. [Accessed 25-Jan-2023]."},{"key":"e_1_3_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-020-0157-6"},{"key":"e_1_3_2_2_47_1","volume-title":"International Journal of Transgender Health","author":"Puckett Jae A","year":"2021","unstructured":"Jae A Puckett, Alix B Aboussouan, Allura L Ralston, Brian Mustanski, and Michael E Newcomb. 2021. Systems of cissexism and the daily production of stress for transgender and gender diverse people. International Journal of Transgender Health (2021), 1\u201314."},{"key":"e_1_3_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.1002\/jclp.22865"},{"key":"e_1_3_2_2_49_1","volume-title":"Rebuilding Trust: Queer in AI Approach to Artificial Intelligence Risk Management. queerinai.org\/risk-management","author":"Organizers","year":"2021","unstructured":"Organizers of QueerInAI, Ashwin S, William Agnew, Hetvi Jethwani, and Arjun Subramonian. 2021. Rebuilding Trust: Queer in AI Approach to Artificial Intelligence Risk Management. queerinai.org\/risk-management"},{"key":"e_1_3_2_2_50_1","volume-title":"Language models are unsupervised multitask learners. OpenAI blog 1, 8","author":"Radford Alec","year":"2019","unstructured":"Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, 2019. Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019), 9."},{"key":"e_1_3_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445914"},{"key":"e_1_3_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.fss.2015.06.019"},{"key":"e_1_3_2_2_53_1","unstructured":"Leonard Richardson. (n.d.). Beautiful Soup: We called him Tortoise because he taught us. \u2014 crummy.com. https:\/\/www.crummy.com\/software\/BeautifulSoup\/. [Accessed 05-Feb-2023]."},{"key":"e_1_3_2_2_54_1","doi-asserted-by":"publisher","DOI":"10.1089\/trgh.2016.0012"},{"key":"e_1_3_2_2_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/3361108"},{"key":"e_1_3_2_2_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN52387.2021.9533924"},{"key":"e_1_3_2_2_57_1","doi-asserted-by":"publisher","DOI":"10.1080\/17470210600973390"},{"key":"e_1_3_2_2_58_1","volume-title":"It\u2019s not just size that matters: Small language models are also few-shot learners. arXiv preprint arXiv:2009.07118","author":"Schick Timo","year":"2020","unstructured":"Timo Schick and Hinrich Sch\u00fctze. 2020. It\u2019s not just size that matters: Small language models are also few-shot learners. arXiv preprint arXiv:2009.07118 (2020)."},{"key":"e_1_3_2_2_59_1","volume-title":"The woman worked as a babysitter: On biases in language generation. arXiv preprint arXiv:1909.01326","author":"Sheng Emily","year":"2019","unstructured":"Emily Sheng, Kai-Wei Chang, Premkumar Natarajan, and Nanyun Peng. 2019. The woman worked as a babysitter: On biases in language generation. arXiv preprint arXiv:1909.01326 (2019)."},{"key":"e_1_3_2_2_60_1","volume-title":"Towards controllable biases in language generation. arXiv preprint arXiv:2005.00268","author":"Sheng Emily","year":"2020","unstructured":"Emily Sheng, Kai-Wei Chang, Premkumar Natarajan, and Nanyun Peng. 2020. Towards controllable biases in language generation. arXiv preprint arXiv:2005.00268 (2020)."},{"key":"e_1_3_2_2_61_1","volume-title":"Societal biases in language generation: Progress and challenges. arXiv preprint arXiv:2105.04054","author":"Sheng Emily","year":"2021","unstructured":"Emily Sheng, Kai-Wei Chang, Premkumar Natarajan, and Nanyun Peng. 2021. Societal biases in language generation: Progress and challenges. arXiv preprint arXiv:2105.04054 (2021)."},{"key":"e_1_3_2_2_62_1","volume-title":"Generic masculine words and thinking. Women\u2019s Studies International Quarterly 3, 2-3","author":"Silveira Jeanette","year":"1980","unstructured":"Jeanette Silveira. 1980. Generic masculine words and thinking. Women\u2019s Studies International Quarterly 3, 2-3 (1980), 165\u2013178."},{"key":"e_1_3_2_2_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376315"},{"key":"e_1_3_2_2_64_1","doi-asserted-by":"publisher","DOI":"10.5749\/j.ctttv2st"},{"key":"e_1_3_2_2_65_1","volume-title":"Development of the gender minority stress and resilience measure.Psychology of Sexual Orientation and Gender Diversity 2, 1","author":"Testa Rylan J","year":"2015","unstructured":"Rylan J Testa, Janice Habarth, Jayme Peta, Kimberly Balsam, and Walter Bockting. 2015. Development of the gender minority stress and resilience measure.Psychology of Sexual Orientation and Gender Diversity 2, 1 (2015), 65."},{"key":"e_1_3_2_2_66_1","volume-title":"Suicidal ideation in transgender people: Gender minority stress and interpersonal theory factors.Journal of abnormal psychology 126, 1","author":"Testa Rylan J","year":"2017","unstructured":"Rylan J Testa, Matthew S Michaels, Whitney Bliss, Megan L Rogers, Kimberly F Balsam, and Thomas Joiner. 2017. Suicidal ideation in transgender people: Gender minority stress and interpersonal theory factors.Journal of abnormal psychology 126, 1 (2017), 125."},{"key":"e_1_3_2_2_67_1","doi-asserted-by":"publisher","DOI":"10.1002\/wcs.1583"},{"key":"e_1_3_2_2_68_1","volume-title":"Kirsty Anderson, Pushmeet Kohli, Ben Coppin, and Po-Sen Huang.","author":"Welbl Johannes","year":"2021","unstructured":"Johannes Welbl, Amelia Glaese, Jonathan Uesato, Sumanth Dathathri, John Mellor, Lisa Anne Hendricks, Kirsty Anderson, Pushmeet Kohli, Ben Coppin, and Po-Sen Huang. 2021. Challenges in detoxifying language models. arXiv preprint arXiv:2109.07445 (2021)."},{"key":"e_1_3_2_2_69_1","volume-title":"Xi Victoria Lin","author":"Zhang Susan","year":"2022","unstructured":"Susan Zhang, Stephen Roller, Naman Goyal, Mikel Artetxe, Moya Chen, Shuohui Chen, Christopher Dewan, Mona Diab, Xian Li, Xi Victoria Lin, 2022. Opt: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068 (2022)."}],"event":{"name":"FAccT '23: the 2023 ACM Conference on Fairness, Accountability, and Transparency","location":"Chicago IL USA","acronym":"FAccT '23"},"container-title":["2023 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594078","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3593013.3594078","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:18Z","timestamp":1750178238000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594078"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,12]]},"references-count":69,"alternative-id":["10.1145\/3593013.3594078","10.1145\/3593013"],"URL":"https:\/\/doi.org\/10.1145\/3593013.3594078","relation":{},"subject":[],"published":{"date-parts":[[2023,6,12]]},"assertion":[{"value":"2023-06-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}