{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T20:52:13Z","timestamp":1776286333433,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":61,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T00:00:00Z","timestamp":1655683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"JST CREST","award":["JPMJCR20D3"],"award-info":[{"award-number":["JPMJCR20D3"]}]},{"name":"JSPS KAKENHI","award":["JP22K12091"],"award-info":[{"award-number":["JP22K12091"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,21]]},"DOI":"10.1145\/3531146.3533184","type":"proceedings-article","created":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T14:27:10Z","timestamp":1655735230000},"page":"1280-1292","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":25,"title":["Gender and Racial Bias in Visual Question Answering Datasets"],"prefix":"10.1145","author":[{"given":"Yusuke","family":"Hirota","sequence":"first","affiliation":[{"name":"Osaka University, Japan"}]},{"given":"Yuta","family":"Nakashima","sequence":"additional","affiliation":[{"name":"Osaka University, Japan"}]},{"given":"Noa","family":"Garcia","sequence":"additional","affiliation":[{"name":"Osaka University, Japan"}]}],"member":"320","published-online":{"date-parts":[[2022,6,20]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"1st Workshop on Language for 3D Scenes. Workshop at CVPR","author":"Achlioptas Panos","year":"2021","unstructured":"Panos Achlioptas, Zhenyu Chen, Mohamed Elhoseiny, Angel\u00a0X Chang, Matthias Niessner, and Leonidas Guibas. 2021. 1st Workshop on Language for 3D Scenes. Workshop at CVPR 2021. https:\/\/language3dscenes.github.io\/."},{"key":"e_1_3_2_1_2_1","volume-title":"Don\u2019t Just Assume","author":"Agrawal Aishwarya","unstructured":"Aishwarya Agrawal, Dhruv Batra, Devi Parikh, and Aniruddha Kembhavi. 2018. Don\u2019t Just Assume; Look and Answer: Overcoming Priors for Visual Question Answering. In CVPR."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"Peter Anderson Xiaodong He Chris Buehler Damien Teney Mark Johnson Stephen Gould and Lei Zhang. 2018. Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering. In CVPR.","DOI":"10.1109\/CVPR.2018.00636"},{"key":"e_1_3_2_1_4_1","volume-title":"VQA: Visual Question Answering. In ICCV.","author":"Antol Stanislaw","year":"2015","unstructured":"Stanislaw Antol, Aishwarya Agrawal, Jiasen Lu, Margaret Mitchell, Dhruv Batra, C\u00a0Lawrence Zitnick, and Devi Parikh. 2015. VQA: Visual Question Answering. In ICCV."},{"key":"e_1_3_2_1_5_1","volume-title":"MUTAN: Multimodal Tucker Fusion for Visual Question Answering. In ICCV.","author":"Ben-Younes Hedi","year":"2017","unstructured":"Hedi Ben-Younes, R\u00e9mi Cadene, Matthieu Cord, and Nicolas Thome. 2017. MUTAN: Multimodal Tucker Fusion for Visual Question Answering. In ICCV."},{"key":"e_1_3_2_1_6_1","volume-title":"Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings. NeurIPS","author":"Bolukbasi Tolga","year":"2016","unstructured":"Tolga Bolukbasi, Kai-Wei Chang, James\u00a0Y Zou, Venkatesh Saligrama, and Adam\u00a0T Kalai. 2016. Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings. NeurIPS (2016)."},{"key":"e_1_3_2_1_7_1","volume-title":"Workshop on Multilingual Multimodal Learning. Workshop at ACL","author":"Bugliarello Emanuele","year":"2021","unstructured":"Emanuele Bugliarello, Kai-Wei Chang, Desmond Elliott, Spandana Gella, Aishwarya Kamath, Liunian\u00a0Harold Li, Fangyu Liu, Jonas Pfeiffer, Edoardo\u00a0M. Ponti, Krishna Srinivasan, Ivan Vuli\u0107, Yinfei Yang, and Da Yin. 2021. Workshop on Multilingual Multimodal Learning. Workshop at ACL 2021. https:\/\/mml-workshop.github.io\/."},{"key":"e_1_3_2_1_8_1","volume-title":"Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification. In ACM FAccT.","author":"Buolamwini Joy","year":"2018","unstructured":"Joy Buolamwini and Timnit Gebru. 2018. Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification. In ACM FAccT."},{"key":"e_1_3_2_1_9_1","unstructured":"Kaylee Burns Lisa\u00a0Anne Hendricks Kate Saenko Trevor Darrell and Anna Rohrbach. 2018. Women Also Snowboard: Overcoming Bias in Captioning Models. In ECCV."},{"key":"e_1_3_2_1_10_1","volume-title":"MUREL: Multimodal Relational Reasoning for Visual Question Answering. In CVPR.","author":"Cadene Remi","year":"2019","unstructured":"Remi Cadene, Hedi Ben-Younes, Matthieu Cord, and Nicolas Thome. 2019. MUREL: Multimodal Relational Reasoning for Visual Question Answering. In CVPR."},{"key":"e_1_3_2_1_11_1","volume-title":"RUBI: Reducing Unimodal Biases for Visual Question Answering. NeurIPS","author":"Cadene Remi","year":"2019","unstructured":"Remi Cadene, Corentin Dancette, Matthieu Cord, Devi Parikh, 2019. RUBI: Reducing Unimodal Biases for Visual Question Answering. NeurIPS (2019)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"crossref","unstructured":"Long Chen Xin Yan Jun Xiao Hanwang Zhang Shiliang Pu and Yueting Zhuang. 2020. Counterfactual Samples Synthesizing for Robust Visual Question Answering. In CVPR.","DOI":"10.1109\/CVPR42600.2020.01081"},{"key":"e_1_3_2_1_13_1","volume-title":"Microsoft COCO Captions: Data Collection and Evaluation Server. ECCV","author":"Chen Xinlei","year":"2014","unstructured":"Xinlei Chen, Hao Fang, Tsung-Yi Lin, Ramakrishna Vedantam, Saurabh Gupta, Piotr Doll\u00e1r, and C\u00a0Lawrence Zitnick. 2014. Microsoft COCO Captions: Data Collection and Evaluation Server. ECCV (2014)."},{"key":"e_1_3_2_1_14_1","volume-title":"UNITER: Universal Image-Text Representation Learning. In ECCV.","author":"Chen Yen-Chun","year":"2020","unstructured":"Yen-Chun Chen, Linjie Li, Licheng Yu, Ahmed El\u00a0Kholy, Faisal Ahmed, Zhe Gan, Yu Cheng, and Jingjing Liu. 2020. UNITER: Universal Image-Text Representation Learning. In ECCV."},{"key":"e_1_3_2_1_15_1","volume-title":"Don\u2019t Take the Easy Way Out: Ensemble Based Methods for Avoiding Known Dataset Biases. EMNLP\/IJCNLP (1)","author":"Clark Christopher","year":"2019","unstructured":"Christopher Clark, Mark Yatskar, and Luke Zettlemoyer. 2019. Don\u2019t Take the Easy Way Out: Ensemble Based Methods for Avoiding Known Dataset Biases. EMNLP\/IJCNLP (1) (2019)."},{"key":"e_1_3_2_1_16_1","volume-title":"4th Workshop on Closing the Loop Between Vision and Language. Workshop at ICCV","author":"Elhoseiny Mohamed","year":"2021","unstructured":"Mohamed Elhoseiny, Xin\u00a0Eric Wang, Andrew Brown, Anna Rohrbach, and Marcus Rohrbach. 2021. 4th Workshop on Closing the Loop Between Vision and Language. Workshop at ICCV 2021. https:\/\/sites.google.com\/view\/iccv21clvl."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6713"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"crossref","unstructured":"Robert Geirhos J\u00f6rn-Henrik Jacobsen Claudio Michaelis Richard Zemel Wieland Brendel Matthias Bethge and Felix\u00a0A Wichmann. 2020. Shortcut Learning in Deep Neural Networks. Nature Machine Intelligence(2020).","DOI":"10.1038\/s42256-020-00257-z"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"crossref","unstructured":"Yash Goyal Tejas Khot Douglas Summers-Stay Dhruv Batra and Devi Parikh. 2017. Making the V in VQA Matter: Elevating the Role of Image Understanding in Visual Question Answering. In CVPR.","DOI":"10.1109\/CVPR.2017.670"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"crossref","unstructured":"Danna Gurari Qing Li Abigale\u00a0J Stangl Anhong Guo Chi Lin Kristen Grauman Jiebo Luo and Jeffrey\u00a0P Bigham. 2018. VizWiz Grand Challenge: Answering Visual Questions from Blind People. In CVPR.","DOI":"10.1109\/CVPR.2018.00380"},{"key":"e_1_3_2_1_21_1","volume-title":"Andrew Smart, and Jamila Smith-Loud","author":"Hanna Alex","year":"2020","unstructured":"Alex Hanna, Emily Denton, Andrew Smart, and Jamila Smith-Loud. 2020. Towards a Critical Race Methodology in Algorithmic Fairness. In ACM FAccT."},{"key":"e_1_3_2_1_22_1","volume-title":"Visual Question Answering with Textual Representations for Images. In ICCV Workshops.","author":"Hirota Yusuke","year":"2021","unstructured":"Yusuke Hirota, Noa Garcia, Mayu Otani, Chenhui Chu, Yuta Nakashima, Ittetsu Taniguchi, and Takao Onoye. 2021. Visual Question Answering with Textual Representations for Images. In ICCV Workshops."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"crossref","unstructured":"Yusuke Hirota Yuta Nakashima and Noa Garcia. 2022. Quantifying Societal Bias Amplification in Image Captioning. In CVPR.","DOI":"10.1109\/CVPR52688.2022.01309"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"crossref","unstructured":"Zhicheng Huang Zhaoyang Zeng Yupan Huang Bei Liu Dongmei Fu and Jianlong Fu. 2021. Seeing Out of tHe bOx: End-to-End Pre-training for Vision-Language Representation Learning. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01278"},{"key":"e_1_3_2_1_25_1","volume-title":"GQA: A New Dataset for Real-World Visual Reasoning and Compositional Question Answering. In CVPR.","author":"Hudson A","year":"2019","unstructured":"Drew\u00a0A Hudson and Christopher\u00a0D Manning. 2019. GQA: A New Dataset for Real-World Visual Reasoning and Compositional Question Answering. In CVPR."},{"key":"e_1_3_2_1_26_1","volume-title":"Mitigating Gender Bias Amplification in Distribution by Posterior Regularization. ACL","author":"Jia Shengyu","year":"2020","unstructured":"Shengyu Jia, Tao Meng, Jieyu Zhao, and Kai-Wei Chang. 2020. Mitigating Gender Bias Amplification in Distribution by Posterior Regularization. ACL (2020)."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"crossref","unstructured":"Huaizu Jiang Ishan Misra Marcus Rohrbach Erik Learned-Miller and Xinlei Chen. 2020. In Defense of Grid Features for Visual Question Answering. In CVPR.","DOI":"10.1109\/CVPR42600.2020.01028"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"crossref","unstructured":"Corentin Kervadec Grigory Antipov Moez Baccouche and Christian Wolf. 2021. Roses Are Red Violets Are Blue... But Should VQA Expect Them To?. In CVPR.","DOI":"10.1109\/CVPR46437.2021.00280"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"crossref","unstructured":"Zaid Khan and Yun Fu. 2021. One Label One Billion Faces: Usage and Consistency of Racial Categories in Computer Vision. In ACM FAccT.","DOI":"10.1145\/3442188.3445920"},{"key":"e_1_3_2_1_30_1","volume-title":"Bilinear Attention Networks. NeurIPS","author":"Kim Jin-Hwa","year":"2018","unstructured":"Jin-Hwa Kim, Jaehyun Jun, and Byoung-Tak Zhang. 2018. Bilinear Attention Networks. NeurIPS (2018)."},{"key":"e_1_3_2_1_31_1","volume-title":"Jordan Cheney, Kristen Allen, Patrick Grother, Alan Mah, and Anil\u00a0K Jain.","author":"Klare F","year":"2015","unstructured":"Brendan\u00a0F Klare, Ben Klein, Emma Taborsky, Austin Blanton, Jordan Cheney, Kristen Allen, Patrick Grother, Alan Mah, and Anil\u00a0K Jain. 2015. Pushing the Frontiers of Unconstrained Face Detection and Recognition: Iarpa Janus Benchmark a. In CVPR."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-016-0981-7"},{"key":"e_1_3_2_1_33_1","unstructured":"Liunian\u00a0Harold Li Mark Yatskar Da Yin Cho-Jui Hsieh and Kai-Wei Chang. 2019. VisualBert: A Simple and Performant Baseline for Vision and Language. arXiv preprint arXiv:1908.03557(2019)."},{"key":"e_1_3_2_1_34_1","volume-title":"OSCAR: Object-Semantics Aligned Pre-training for Vision-Language Tasks. In ECCV.","author":"Li Xiujun","year":"2020","unstructured":"Xiujun Li, Xi Yin, Chunyuan Li, Pengchuan Zhang, Xiaowei Hu, Lei Zhang, Lijuan Wang, Houdong Hu, Li Dong, Furu Wei, 2020. OSCAR: Object-Semantics Aligned Pre-training for Vision-Language Tasks. In ECCV."},{"key":"e_1_3_2_1_35_1","unstructured":"Tsung-Yi Lin Michael Maire Serge Belongie James Hays Pietro Perona Deva Ramanan Piotr Doll\u00e1r and C\u00a0Lawrence Zitnick. 2014. Microsoft COCO: Common Objects in Context. In ECCV."},{"key":"e_1_3_2_1_36_1","volume-title":"VilBert: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks. NeurIPS","author":"Lu Jiasen","year":"2019","unstructured":"Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan Lee. 2019. VilBert: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks. NeurIPS (2019)."},{"key":"e_1_3_2_1_37_1","volume-title":"A Multi-World Approach to Question Answering about Real-World Scenes based on Uncertain Input. NeurIPS","author":"Malinowski Mateusz","year":"2014","unstructured":"Mateusz Malinowski and Mario Fritz. 2014. A Multi-World Approach to Question Answering about Real-World Scenes based on Uncertain Input. NeurIPS (2014)."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"crossref","unstructured":"Varun Manjunatha Nirat Saini and Larry\u00a0S Davis. 2019. Explicit Bias Discovery in Visual Question Answering Models. In CVPR.","DOI":"10.1109\/CVPR.2019.00979"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"crossref","unstructured":"Kenneth Marino Mohammad Rastegari Ali Farhadi and Roozbeh Mottaghi. 2019. OK-VQA: A Visual Question Answering Benchmark Requiring External Knowledge. In CVPR.","DOI":"10.1109\/CVPR.2019.00331"},{"key":"e_1_3_2_1_40_1","volume-title":"LANTERN - The Third Workshop Beyond Vision and Language: Integrating Real World Knowledge. Workshop at EACL 2021","author":"Mosbach Marius","year":"2021","unstructured":"Marius Mosbach, Sandro Pezzelle, Michael\u00a0A. Hedderich, Dietrich Klakow, Marie-Francine Moens, and Zeynep Akata. 2021. LANTERN - The Third Workshop Beyond Vision and Language: Integrating Real World Knowledge. Workshop at EACL 2021. https:\/\/www.lantern.uni-saarland.de\/2021\/."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"crossref","unstructured":"Yulei Niu Kaihua Tang Hanwang Zhang Zhiwu Lu Xian-Sheng Hua and Ji-Rong Wen. 2021. Counterfactual VQA: A Cause-Effect Look at Language Bias. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01251"},{"key":"e_1_3_2_1_42_1","unstructured":"Madison Park. 2016. U.S. government to stop using these words to refer to minorities. CNN. https:\/\/edition.cnn.com\/2016\/05\/22\/politics\/obama-federal-law-minorities-references\/index.html."},{"key":"e_1_3_2_1_43_1","volume-title":"No Classification Without Representation: Assessing Geodiversity Issues in Open Data Sets for the Developing World. NeurIPS Workshop","author":"Shankar Shreya","year":"2017","unstructured":"Shreya Shankar, Yoni Halpern, Eric Breck, James Atwood, Jimbo Wilson, and D Sculley. 2017. No Classification Without Representation: Assessing Geodiversity Issues in Open Data Sets for the Developing World. NeurIPS Workshop (2017)."},{"key":"e_1_3_2_1_44_1","volume-title":"Conceptual Captions: A Cleaned, Hypernymed, Image Alt-text Dataset for Automatic Image Captioning. In ACL.","author":"Sharma Piyush","year":"2018","unstructured":"Piyush Sharma, Nan Ding, Sebastian Goodman, and Radu Soricut. 2018. Conceptual Captions: A Cleaned, Hypernymed, Image Alt-text Dataset for Automatic Image Captioning. In ACL."},{"key":"e_1_3_2_1_45_1","volume-title":"Visual Question Answering Workshop. Workshop at CVPR","author":"Shrivastava Ayush","year":"2021","unstructured":"Ayush Shrivastava, Yash\u00a0Mukund Kant, Satwik Kottur, Dhruv Batra, Devi Parikh, and Aishwarya Agrawal. 2021. Visual Question Answering Workshop. Workshop at CVPR 2021. https:\/\/visualqa.org\/workshop."},{"key":"e_1_3_2_1_46_1","volume-title":"LXMERT: Learning Cross-Modality Encoder Representations from Transformers. EMNLP\/IJCNLP","author":"Tan Hao","year":"2019","unstructured":"Hao Tan and Mohit Bansal. 2019. LXMERT: Learning Cross-Modality Encoder Representations from Transformers. EMNLP\/IJCNLP (2019)."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"crossref","unstructured":"Ruixiang Tang Mengnan Du Yuening Li Zirui Liu Na Zou and Xia Hu. 2021. Mitigating Gender Bias in Captioning Systems. In WWW.","DOI":"10.1145\/3442381.3449950"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/2812802"},{"key":"e_1_3_2_1_49_1","volume-title":"Feature and Label Embedding Spaces Matter in Addressing Image Classifier Bias. BMVC","author":"Thong William","year":"2021","unstructured":"William Thong and Cees\u00a0GM Snoek. 2021. Feature and Label Embedding Spaces Matter in Addressing Image Classifier Bias. BMVC (2021)."},{"key":"e_1_3_2_1_50_1","volume-title":"Boys","author":"Vleuten Maaike Van\u00a0der","year":"2016","unstructured":"Maaike Van\u00a0der Vleuten, Eva Jaspers, Ineke Maas, and Tanja van\u00a0der Lippe. 2016. Boys\u2019 and Girls\u2019 Educational Choices in Secondary Education. The Role of Gender Ideology. Educational Studies (2016)."},{"key":"e_1_3_2_1_51_1","volume-title":"Show and Tell: Lessons Learned from the 2015 MSCOCO Image Captioning Challenge. Trans.\u00a0PAMI","author":"Vinyals Oriol","year":"2016","unstructured":"Oriol Vinyals, Alexander Toshev, Samy Bengio, and Dumitru Erhan. 2016. Show and Tell: Lessons Learned from the 2015 MSCOCO Image Captioning Challenge. Trans.\u00a0PAMI (2016)."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"crossref","unstructured":"Mei Wang Weihong Deng Jiani Hu Xunqiang Tao and Yaohai Huang. 2019. Racial Faces in the Wild: Reducing Racial Bias by Information Maximization Adaptation Network. In ICCV.","DOI":"10.1109\/ICCV.2019.00078"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"crossref","unstructured":"Tan Wang Jianqiang Huang Hanwang Zhang and Qianru Sun. 2020. Visual Commonsense R-CNN. In CVPR.","DOI":"10.1109\/CVPR42600.2020.01077"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"crossref","unstructured":"Tianlu Wang Jieyu Zhao Mark Yatskar Kai-Wei Chang and Vicente Ordonez. 2019. Balanced Datasets are Not Enough: Estimating and Mitigating Gender Bias in Deep Image Representations. In ICCV.","DOI":"10.1109\/ICCV.2019.00541"},{"key":"e_1_3_2_1_55_1","volume-title":"From VQA to VLN: Recent Advances in Vision-and-Language Research. Tutorial at CVPR","author":"Wu Qi","year":"2021","unstructured":"Qi Wu and Zhe Gan. 2021. From VQA to VLN: Recent Advances in Vision-and-Language Research. Tutorial at CVPR 2021. https:\/\/vqa2vln-tutorial.github.io\/."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"crossref","unstructured":"Zekun Yang Noa Garcia Chenhui Chu Mayu Otani Yuta Nakashima and Haruo Takemura. 2020. BERT Representations for Video Question Answering. In WACV.","DOI":"10.1109\/WACV45572.2020.9093596"},{"key":"e_1_3_2_1_57_1","unstructured":"Quanzeng You Hailin Jin Zhaowen Wang Chen Fang and Jiebo Luo. 2016. Image Captioning with Semantic Attention. In CVPR."},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"crossref","unstructured":"Pengchuan Zhang Xiujun Li Xiaowei Hu Jianwei Yang Lei Zhang Lijuan Wang Yejin Choi and Jianfeng Gao. 2021. VinVL: Revisiting Visual Representations in Vision-Language Models. In CVPR.","DOI":"10.1109\/CVPR46437.2021.00553"},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"crossref","unstructured":"Dora Zhao Angelina Wang and Olga Russakovsky. 2021. Understanding and Evaluating Racial Biases in Image Captioning. In ICCV.","DOI":"10.1109\/ICCV48922.2021.01456"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"crossref","unstructured":"Jieyu Zhao Tianlu Wang Mark Yatskar Vicente Ordonez and Kai-Wei Chang. 2017. Men Also Like Shopping: Reducing Gender Bias Amplification using Corpus-level Constraints. In EMNLP.","DOI":"10.18653\/v1\/D17-1323"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"crossref","unstructured":"Yuke Zhu Oliver Groth Michael Bernstein and Li Fei-Fei. 2016. Visual7W: Grounded Question Answering in Images. In CVPR.","DOI":"10.1109\/CVPR.2016.540"}],"event":{"name":"FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency","location":"Seoul Republic of Korea","acronym":"FAccT '22","sponsor":["ACM Association for Computing Machinery"]},"container-title":["2022 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533184","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3533184","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:02:09Z","timestamp":1750186929000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533184"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,20]]},"references-count":61,"alternative-id":["10.1145\/3531146.3533184","10.1145\/3531146"],"URL":"https:\/\/doi.org\/10.1145\/3531146.3533184","relation":{},"subject":[],"published":{"date-parts":[[2022,6,20]]},"assertion":[{"value":"2022-06-20","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}