{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,18]],"date-time":"2026-01-18T07:46:08Z","timestamp":1768722368474,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":101,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T00:00:00Z","timestamp":1655683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["IIS-1910064 and IIS-1553088"],"award-info":[{"award-number":["IIS-1910064 and IIS-1553088"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Sloan Foundation","award":["Sloan Fellowship"],"award-info":[{"award-number":["Sloan Fellowship"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,21]]},"DOI":"10.1145\/3531146.3533128","type":"proceedings-article","created":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T14:27:10Z","timestamp":1655735230000},"page":"637-650","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":8,"title":["Subverting Fair Image Search with Generative Adversarial Perturbations"],"prefix":"10.1145","author":[{"given":"Avijit","family":"Ghosh","sequence":"first","affiliation":[{"name":"Khoury College of Computer Sciences, Northeastern University, USA"}]},{"given":"Matthew","family":"Jagielski","sequence":"additional","affiliation":[{"name":"Google Brain, USA"}]},{"given":"Christo","family":"Wilson","sequence":"additional","affiliation":[{"name":"Khoury College of Computer Sciences, Northeastern University, USA"}]}],"member":"320","published-online":{"date-parts":[[2022,6,20]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Congress (2019-2020)","year":"2019","unstructured":"116th Congress (2019-2020). [n.d.]. H.R.2231 - Algorithmic Accountability Act of 2019. https:\/\/www.congress.gov\/bill\/116th-congress\/house-bill\/2231."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"Dzifa Adjaye-Gbewonyo Robert\u00a0A Bednarczyk Robert\u00a0L Davis and Saad\u00a0B Omer. 2014. Using the Bayesian Improved Surname Geocoding Method (BISG) to create a working classification of race and ethnicity in a diverse managed care population: a validation study. Health services research 49 1 (2014) 268\u2013283.","DOI":"10.1111\/1475-6773.12089"},{"key":"e_1_3_2_1_3_1","unstructured":"Alekh Agarwal Miroslav Dud\u00edk and Zhiwei\u00a0Steven Wu. 2019. Fair regression: Quantitative definitions and reduction-based algorithms. arXiv preprint arXiv:1905.12843(2019)."},{"key":"e_1_3_2_1_4_1","unstructured":"Facebook AI. 2021. How we\u2019re using Fairness Flow to help build AI that works better for everyone. Facebook AI. https:\/\/ai.facebook.com\/blog\/how-were-using-fairness-flow-to-help-build-ai-that-works-better-for-everyone\/."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00357"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445888"},{"key":"e_1_3_2_1_7_1","volume-title":"Machine bias: There\u2019s software used across the country to predict future criminals and it\u2019s biased against blacks","author":"Angwin Julia","year":"2016","unstructured":"Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. 2019. Machine bias: There\u2019s software used across the country to predict future criminals and it\u2019s biased against blacks. 2016. URL https:\/\/www. propublica. org\/article\/machine-bias-risk-assessments-in-criminal-sentencing(2019)."},{"key":"e_1_3_2_1_8_1","volume-title":"International Conference on Machine Learning. PMLR, 274\u2013283","author":"Athalye Anish","year":"2018","unstructured":"Anish Athalye, Nicholas Carlini, and David Wagner. 2018. Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples. In International Conference on Machine Learning. PMLR, 274\u2013283."},{"key":"e_1_3_2_1_9_1","first-page":"671","article-title":"Big data\u2019s disparate impact","volume":"104","author":"Barocas Solon","year":"2016","unstructured":"Solon Barocas and Andrew\u00a0D Selbst. 2016. Big data\u2019s disparate impact. Calif. L. Rev. 104(2016), 671.","journal-title":"Calif. L. Rev."},{"key":"e_1_3_2_1_10_1","unstructured":"Sid Basu Ruthie Berman Adam Bloomston John Campbell Anne Diaz Nanako Era Benjamin Evans Sukhada Palkar and Skyler Wharton. 2020. Measuring discrepancies in Airbnb guest acceptance rates using anonymized demographic data. AirBNB. https:\/\/news.airbnb.com\/wp-content\/uploads\/sites\/4\/2020\/06\/Project-Lighthouse-Airbnb-2020-06-12.pdf."},{"key":"e_1_3_2_1_11_1","volume-title":"Sex and credit: Is there a gender bias in lending?Journal of Banking and Finance 87","author":"Beck Thorsten","year":"2018","unstructured":"Thorsten Beck, Patrick Behr, and Andreas Madestam. 2018. Sex and credit: Is there a gender bias in lending?Journal of Banking and Finance 87 (2018)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8682430"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1147\/JRD.2019.2942287"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445498"},{"key":"e_1_3_2_1_15_1","unstructured":"Richard Berk Hoda Heidari Shahin Jabbari Matthew Joseph Michael Kearns Jamie Morgenstern Seth Neel and Aaron Roth. 2017. A convex framework for fair regression. arXiv preprint arXiv:1706.02409(2017)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"crossref","unstructured":"Alex Beutel Jilin Chen Tulsee Doshi Hai Qian Li Wei Yi Wu Lukasz Heldt Zhe Zhao Lichan Hong Ed\u00a0H. Chi and Cristos Goodrow. 2019. Fairness in Recommendation Ranking through Pairwise Comparisons. In KDD. https:\/\/arxiv.org\/pdf\/1903.00780.pdf","DOI":"10.1145\/3292500.3330745"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372877"},{"key":"e_1_3_2_1_18_1","unstructured":"Tolga Bolukbasi Kai-Wei Chang James\u00a0Y Zou Venkatesh Saligrama and Adam\u00a0T Kalai. 2016. Man is to computer programmer as woman is to homemaker? debiasing word embeddings. In Advances in neural information processing systems. 4349\u20134357."},{"key":"e_1_3_2_1_19_1","unstructured":"Tom\u00a0B Brown Dandelion Man\u00e9 Aurko Roy Mart\u00edn Abadi and Justin Gilmer. 2017. Adversarial patch. arXiv preprint arXiv:1712.09665(2017)."},{"key":"e_1_3_2_1_20_1","volume-title":"International Conference on Machine Learning. 803\u2013811","author":"Brunet Marc-Etienne","year":"2019","unstructured":"Marc-Etienne Brunet, Colleen Alkalay-Houlihan, Ashton Anderson, and Richard Zemel. 2019. Understanding the origins of bias in word embeddings. In International Conference on Machine Learning. 803\u2013811."},{"key":"e_1_3_2_1_21_1","volume-title":"Conference on fairness, accountability and transparency. PMLR, 77\u201391","author":"Buolamwini Joy","year":"2018","unstructured":"Joy Buolamwini and Timnit Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency. PMLR, 77\u201391."},{"key":"e_1_3_2_1_22_1","unstructured":"Consumer Financial\u00a0Protection Bureau. 2014. Using publicly available information to proxy for unidentified race and ethnicity. Report available at https:\/\/files.consumerfinance.gov\/f\/201409_cfpb_report_proxy-methodology.pdf(2014)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/290941.291025"},{"key":"e_1_3_2_1_24_1","volume-title":"Towards evaluating the robustness of neural networks. In 2017 ieee symposium on security and privacy (sp)","author":"Carlini Nicholas","unstructured":"Nicholas Carlini and David Wagner. 2017. Towards evaluating the robustness of neural networks. In 2017 ieee symposium on security and privacy (sp). IEEE, 39\u201357."},{"key":"e_1_3_2_1_25_1","volume-title":"International Conference on Machine Learning. PMLR, 1349\u20131361","author":"Celis L\u00a0Elisa","year":"2021","unstructured":"L\u00a0Elisa Celis, Lingxiao Huang, Vijay Keswani, and Nisheeth\u00a0K Vishnoi. 2021. Fair classification with noisy protected attributes: A framework with provable guarantees. In International Conference on Machine Learning. PMLR, 1349\u20131361."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3415210"},{"key":"e_1_3_2_1_27_1","unstructured":"L\u00a0Elisa Celis Anay Mehrotra and Nisheeth\u00a0K Vishnoi. 2021. Fair Classification with Adversarial Perturbations. arXiv preprint arXiv:2106.05964(2021)."},{"key":"e_1_3_2_1_28_1","volume-title":"45th International Colloquium on Automata, Languages, and Programming (ICALP","author":"Celis L\u00a0Elisa","year":"2018","unstructured":"L\u00a0Elisa Celis, Damian Straszak, and Nisheeth\u00a0K Vishnoi. 2018. Ranking with Fairness Constraints. In 45th International Colloquium on Automata, Languages, and Programming (ICALP 2018). Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik."},{"key":"e_1_3_2_1_29_1","unstructured":"Hongyan Chang Ta\u00a0Duy Nguyen Sasi\u00a0Kumar Murakonda Ehsan Kazemi and Reza Shokri. 2020. On Adversarial Bias and the Robustness of Fair Machine Learning. arXiv preprint arXiv:2006.08669(2020)."},{"key":"e_1_3_2_1_30_1","unstructured":"European Commission. [n.d.]. Proposal for a Regulation laying down harmonised rules on artificial intelligence (Artificial Intelligence Act). https:\/\/digital-strategy.ec.europa.eu\/en\/library\/proposal-regulation-laying-down-harmonised-rules-artificial-intelligence-artificial-intelligence."},{"key":"e_1_3_2_1_31_1","volume-title":"International conference on machine learning. PMLR, 1115\u20131124","author":"Dai Hanjun","year":"2018","unstructured":"Hanjun Dai, Hui Li, Tian Tian, Xin Huang, Lin Wang, Jun Zhu, and Le Song. 2018. Adversarial attack on graph structured data. In International conference on machine learning. PMLR, 1115\u20131124."},{"key":"e_1_3_2_1_32_1","volume-title":"The Security Cards: A Security Threat Brainstorming Toolkit","author":"Denning Tamara","unstructured":"Tamara Denning, Batya Friedman, and Tadayoshi Kohno. 2013. The Security Cards: A Security Threat Brainstorming Toolkit. University of Washington. https:\/\/securitycards.cs.washington.edu\/."},{"key":"e_1_3_2_1_33_1","unstructured":"UK\u00a0Office for Artificial\u00a0Intelligence. [n.d.]. Ethics Transparency and Accountability Framework for Automated Decision-Making. https:\/\/www.gov.uk\/government\/publications\/ethics-transparency-and-accountability-framework-for-automated-decision-making."},{"key":"e_1_3_2_1_34_1","volume-title":"Value sensitive design: shaping technology with moral imagination","author":"Friedman Batya","unstructured":"Batya Friedman and David Hendry. 2019. Value sensitive design: shaping technology with moral imagination. MIT Press."},{"key":"e_1_3_2_1_35_1","volume-title":"Domain-adversarial training of neural networks. The journal of machine learning research 17, 1","author":"Ganin Yaroslav","year":"2016","unstructured":"Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, Fran\u00e7ois Laviolette, Mario Marchand, and Victor Lempitsky. 2016. Domain-adversarial training of neural networks. The journal of machine learning research 17, 1 (2016), 2096\u20132030."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1720347115"},{"key":"e_1_3_2_1_37_1","volume-title":"Retrieve Fast","author":"Geigle Gregor","year":"1920","unstructured":"Gregor Geigle, Jonas Pfeiffer, Nils Reimers, Ivan Vuli\u0107, and Iryna Gurevych. 2021. Retrieve Fast, Rerank Smart: Cooperative and Joint Approaches for Improved Cross-Modal Retrieval. arXiv preprint abs\/2103.11920 (2021). arxiv:2103.11920http:\/\/arxiv.org\/abs\/2103.11920"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330691"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462850"},{"key":"e_1_3_2_1_40_1","volume-title":"Proceedings of 2nd Workshop on Diversity in Artificial Intelligence (AIDBEI)(Proceedings of Machine Learning Research, Vol.\u00a0142)","author":"Ghosh Avijit","year":"2021","unstructured":"Avijit Ghosh, Lea Genuit, and Mary Reagan. 2021. Characterizing Intersectional Group Fairness with Worst-Case Comparisons. In Proceedings of 2nd Workshop on Diversity in Artificial Intelligence (AIDBEI)(Proceedings of Machine Learning Research, Vol.\u00a0142), Deepti Lamba and William\u00a0H. Hsu (Eds.). PMLR, 22\u201334. https:\/\/proceedings.mlr.press\/v142\/ghosh21a.html"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/SIEDS49339.2020.9106661"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3278721.3278722"},{"key":"e_1_3_2_1_43_1","unstructured":"Ian\u00a0J Goodfellow Jonathon Shlens and Christian Szegedy. 2014. Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572(2014)."},{"key":"e_1_3_2_1_44_1","unstructured":"Danny Goodwin. 2011. Top Google Result Gets 36.4% of Clicks [Study]. Search Engine Watch. https:\/\/www.searchenginewatch.com\/2011\/04\/21\/top-google-result-gets-36-4-of-clicks-study\/."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3209981"},{"key":"e_1_3_2_1_47_1","unstructured":"Lingxiao Huang and Nisheeth\u00a0K Vishnoi. 2019. Stable and fair classification. arXiv preprint arXiv:1902.07823(2019)."},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"crossref","unstructured":"Matthew Jagielski Giorgio Severi Niklas\u00a0Pousette Harger and Alina Oprea. 2020. Subpopulation data poisoning attacks. arXiv preprint arXiv:2006.14026(2020).","DOI":"10.1145\/3460120.3485368"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/582415.582418"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-33486-3_3"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3213586.3226206"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV48630.2021.00159"},{"key":"e_1_3_2_1_53_1","unstructured":"Alistair Knott. [n.d.]. Moving Towards Responsible Government Use of AI in New Zealand). https:\/\/digitaltechitp.nz\/2021\/03\/22\/moving-towards-responsible-government-use-of-ai-in-new-zealand\/."},{"key":"e_1_3_2_1_54_1","unstructured":"Alexey Kurakin Ian Goodfellow and Samy Bengio. 2016. Adversarial machine learning at scale. arXiv preprint arXiv:1611.01236(2016)."},{"key":"e_1_3_2_1_55_1","unstructured":"Preethi Lahoti Alex Beutel Jilin Chen Kang Lee Flavien Prost Nithum Thain Xuezhi Wang and Ed\u00a0H Chi. 2020. Fairness without Demographics through Adversarially Reweighted Learning. (2020)."},{"key":"e_1_3_2_1_56_1","unstructured":"Kristina Lerman Anon Plangprasopchok and Chio Wong. 2007. Personalizing image search results on flickr. Intelligent Information Personalization(2007)."},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00500"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58577-8_8"},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"e_1_3_2_1_60_1","unstructured":"Yanpei Liu Xinyun Chen Chang Liu and Dawn Song. 2016. Delving into transferable adversarial examples and black-box attacks. arXiv preprint arXiv:1611.02770(2016)."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3323873.3325052"},{"key":"e_1_3_2_1_62_1","unstructured":"Joshua\u00a0R Loftus Chris Russell Matt\u00a0J Kusner and Ricardo Silva. 2018. Causal reasoning for algorithmic fairness. arXiv preprint arXiv:1805.05859(2018)."},{"key":"e_1_3_2_1_63_1","volume-title":"Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. arXiv preprint arXiv:1908.02265(2019).","author":"Lu Jiasen","year":"2019","unstructured":"Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan Lee. 2019. Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. arXiv preprint arXiv:1908.02265(2019)."},{"key":"e_1_3_2_1_64_1","unstructured":"Ninareh Mehrabi Muhammad Naveed Fred Morstatter and Aram Galstyan. 2020. Exacerbating Algorithmic Bias through Fairness Attacks. arXiv preprint arXiv:2012.08723(2020)."},{"key":"e_1_3_2_1_65_1","volume-title":"Conference on Fairness, Accountability and Transparency. 107\u2013118","author":"Menon Aditya\u00a0Krishna","year":"2018","unstructured":"Aditya\u00a0Krishna Menon and Robert\u00a0C Williamson. 2018. The cost of fairness in binary classification. In Conference on Fairness, Accountability and Transparency. 107\u2013118."},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"crossref","unstructured":"Seyed-Mohsen Moosavi-Dezfooli Alhussein Fawzi Omar Fawzi and Pascal Frossard. 2017. Universal adversarial perturbations. arxiv:1610.08401\u00a0[cs.CV]","DOI":"10.1109\/CVPR.2017.17"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.282"},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"crossref","unstructured":"Marco Morik Ashudeep Singh Jessica Hong and Thorsten Joachims. 2020. Controlling Fairness and Bias in Dynamic Learning-to-Rank. arXiv preprint arXiv:2005.14713(2020).","DOI":"10.1145\/3397271.3401100"},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-15719-7_23"},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11553"},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445910"},{"key":"e_1_3_2_1_72_1","unstructured":"Jakob Nielsen. 2003. Usability 101: introduction to usability. Jakob Nielsen\u2019s Alertbox."},{"key":"e_1_3_2_1_73_1","unstructured":"Government of Canada. [n.d.]. Responsible use of artificial intelligence (AI). https:\/\/www.canada.ca\/en\/government\/system\/digital-government\/digital-government-innovations\/responsible-use-ai.html."},{"key":"e_1_3_2_1_74_1","unstructured":"Nicolas Papernot Fartash Faghri Nicholas Carlini Ian Goodfellow Reuben Feinman Alexey Kurakin Cihang Xie Yash Sharma Tom Brown Aurko Roy 2016. Technical report on the cleverhans v2. 1.0 adversarial examples library. arXiv preprint arXiv:1610.00768(2016)."},{"key":"e_1_3_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00465"},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372873"},{"key":"e_1_3_2_1_77_1","unstructured":"Nisarg Raval and Manisha Verma. 2020. One word at a time: adversarial attacks on retrieval models. arXiv preprint arXiv:2008.02197(2020)."},{"key":"e_1_3_2_1_78_1","volume-title":"Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems 28","author":"Ren Shaoqing","year":"2015","unstructured":"Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. 2015. Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems 28 (2015), 91\u201399."},{"key":"e_1_3_2_1_79_1","doi-asserted-by":"crossref","unstructured":"Alexey Romanov Maria De-Arteaga Hanna Wallach Jennifer Chayes Christian Borgs Alexandra Chouldechova Sahin Geyik Krishnaram Kenthapadi Anna Rumshisky and Adam\u00a0Tauman Kalai. 2019. What\u2019s in a Name? Reducing Bias in Bios without Access to Protected Attributes. arXiv preprint arXiv:1904.05233(2019).","DOI":"10.18653\/v1\/N19-1424"},{"key":"e_1_3_2_1_80_1","volume-title":"Quantifying the Impact of User Attentionon Fair Group Representation in Ranked Lists. In Companion Proceedings of The 2019 World Wide Web Conference. 553\u2013562","author":"Sapiezynski Piotr","year":"2019","unstructured":"Piotr Sapiezynski, Wesley Zeng, Ronald E\u00a0Robertson, Alan Mislove, and Christo Wilson. 2019. Quantifying the Impact of User Attentionon Fair Group Representation in Ranked Lists. In Companion Proceedings of The 2019 World Wide Web Conference. 553\u2013562."},{"key":"e_1_3_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1145\/3392866"},{"key":"e_1_3_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1109\/ASYU50717.2020.9259802"},{"key":"e_1_3_2_1_83_1","unstructured":"Ali Shafahi W\u00a0Ronny Huang Mahyar Najibi Octavian Suciu Christoph Studer Tudor Dumitras and Tom Goldstein. 2018. Poison frogs! targeted clean-label poisoning attacks on neural networks. arXiv preprint arXiv:1804.00792(2018)."},{"key":"e_1_3_2_1_84_1","unstructured":"Ali Shafahi Mahyar Najibi Amin Ghiasi Zheng Xu John Dickerson Christoph Studer Larry\u00a0S Davis Gavin Taylor and Tom Goldstein. 2019. Adversarial training for free!arXiv preprint arXiv:1904.12843(2019)."},{"key":"e_1_3_2_1_85_1","volume-title":"29th USENIX Security Symposium (USENIX Security 20)","author":"Shan Shawn","year":"2020","unstructured":"Shawn Shan, Emily Wenger, Jiayun Zhang, Huiying Li, Haitao Zheng, and Ben\u00a0Y Zhao. 2020. Fawkes: Protecting privacy against unauthorized deep learning models. In 29th USENIX Security Symposium (USENIX Security 20). 1589\u20131604."},{"key":"e_1_3_2_1_86_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220088"},{"key":"e_1_3_2_1_87_1","doi-asserted-by":"crossref","unstructured":"David Solans Battista Biggio and Carlos Castillo. 2020. Poisoning Attacks on Algorithmic Fairness. arXiv preprint arXiv:2004.07401(2020).","DOI":"10.1007\/978-3-030-67658-2_10"},{"key":"e_1_3_2_1_88_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"e_1_3_2_1_89_1","unstructured":"Christian Szegedy Wojciech Zaremba Ilya Sutskever Joan Bruna Dumitru Erhan Ian Goodfellow and Rob Fergus. 2013. Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199(2013)."},{"key":"e_1_3_2_1_90_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.220"},{"key":"e_1_3_2_1_91_1","unstructured":"Florian Tramer Nicholas Carlini Wieland Brendel and Aleksander Madry. 2020. On adaptive attacks to adversarial example defenses. arXiv preprint arXiv:2002.08347(2020)."},{"key":"e_1_3_2_1_92_1","volume-title":"The Space of Transferable Adversarial Examples. arXiv","author":"Tram\u00e8r Florian","year":"2017","unstructured":"Florian Tram\u00e8r, Nicolas Papernot, Ian Goodfellow, Dan Boneh, and Patrick McDaniel. 2017. The Space of Transferable Adversarial Examples. arXiv (2017). https:\/\/arxiv.org\/abs\/1704.03453"},{"key":"e_1_3_2_1_93_1","unstructured":"Alexander Turner Dimitris Tsipras and Aleksander Madry. 2018. Clean-label backdoor attacks. (2018)."},{"key":"e_1_3_2_1_94_1","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3412705"},{"key":"e_1_3_2_1_95_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-01580-9"},{"key":"e_1_3_2_1_96_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445928"},{"key":"e_1_3_2_1_97_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58548-8_1"},{"key":"e_1_3_2_1_98_1","doi-asserted-by":"publisher","DOI":"10.1145\/3132847.3132938"},{"key":"e_1_3_2_1_99_1","doi-asserted-by":"publisher","DOI":"10.1145\/3366424.3380048"},{"key":"e_1_3_2_1_100_1","volume-title":"Understanding and Evaluating Racial Biases in Image Captioning. In International Conference on Computer Vision (ICCV).","author":"Zhao Dora","year":"2021","unstructured":"Dora Zhao, Angelina Wang, and Olga Russakovsky. 2021. Understanding and Evaluating Racial Biases in Image Captioning. In International Conference on Computer Vision (ICCV)."},{"key":"e_1_3_2_1_101_1","doi-asserted-by":"crossref","unstructured":"Mo Zhou Zhenxing Niu Le Wang Qilin Zhang and Gang Hua. 2020. Adversarial Ranking Attack and Defense. arXiv preprint arXiv:2002.11293(2020).","DOI":"10.1007\/978-3-030-58568-6_46"}],"event":{"name":"FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency","location":"Seoul Republic of Korea","acronym":"FAccT '22","sponsor":["ACM Association for Computing Machinery"]},"container-title":["2022 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533128","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3533128","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3533128","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:02:08Z","timestamp":1750186928000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533128"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,20]]},"references-count":101,"alternative-id":["10.1145\/3531146.3533128","10.1145\/3531146"],"URL":"https:\/\/doi.org\/10.1145\/3531146.3533128","relation":{},"subject":[],"published":{"date-parts":[[2022,6,20]]},"assertion":[{"value":"2022-06-20","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}