{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,17]],"date-time":"2025-10-17T02:40:54Z","timestamp":1760668854721,"version":"build-2065373602"},"reference-count":85,"publisher":"Association for Computing Machinery (ACM)","issue":"7","funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2022YFB3103700, 2022YFB3103701"],"award-info":[{"award-number":["2022YFB3103700, 2022YFB3103701"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62472409, U21B2046, 62272125"],"award-info":[{"award-number":["62472409, U21B2046, 62272125"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Hum.-Comput. Interact."],"published-print":{"date-parts":[[2025,10,18]]},"abstract":"<jats:p>Artificial intelligence (AI) systems are playing an increasingly crucial role in people's lives, and their frequent unfair behaviors raise concerns about fairness. To unveil the unfairness in AI systems, researchers conduct fairness auditing on these systems. However, existing fairness auditing works often focus on group fairness while ignoring discriminatory phenomena among individuals. To unearth discriminatory phenomena against individuals within AI systems, this paper proposes an individual fairness auditing framework, termed ''substantiating'', which can identify discrimination instances within AI systems by constructing individual samples. To construct these samples for substantiating, auditors often have to rely on subjective prior knowledge, lacking guidelines on how to construct unfair samples. To address this issue, this paper introduces two categories of automated sample generation methods that can rapidly find unfair samples within a limited number of queries to the system and demonstrate their effectiveness through experiments. This paper evaluates the proposed auditing framework among three categories of stakeholders in AI fairness: auditors, AI model developers, and non-technical personnel. The research findings point out their demand for individual fairness audits of AI systems and highlight how the framework supports a reliable and convenient individual fairness audit.<\/jats:p>","DOI":"10.1145\/3757414","type":"journal-article","created":{"date-parts":[[2025,10,16]],"date-time":"2025-10-16T16:59:10Z","timestamp":1760633950000},"page":"1-38","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["'I Know You Are Discriminatory!': Automated Substantiating for Individual Fairness Auditing of AI Systems"],"prefix":"10.1145","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-5607-784X","authenticated-orcid":false,"given":"Yuanhao","family":"Liu","sequence":"first","affiliation":[{"name":"State Key Laboratory of AI Safety, Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China and University of Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3454-4789","authenticated-orcid":false,"given":"Qi","family":"Cao","sequence":"additional","affiliation":[{"name":"State Key Laboratory of AI Safety, Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1081-8119","authenticated-orcid":false,"given":"Huawei","family":"Shen","sequence":"additional","affiliation":[{"name":"State Key Laboratory of AI Safety, Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1197-5212","authenticated-orcid":false,"given":"Kaike","family":"Zhang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of AI Safety, Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6994-6791","authenticated-orcid":false,"given":"Yunfan","family":"Wu","sequence":"additional","affiliation":[{"name":"State Key Laboratory of AI Safety, Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5201-8195","authenticated-orcid":false,"given":"Xueqi","family":"Cheng","sequence":"additional","affiliation":[{"name":"State Key Laboratory of AI Safety, Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2025,10,16]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1207\/S15327051HCI1523_5"},{"key":"e_1_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287588"},{"key":"e_1_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359301"},{"key":"e_1_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58592-1_29"},{"key":"e_1_2_1_5_1","unstructured":"Julia Angwin Jeff Larson Lauren Kirchner and Surya Mattu. 2016. Machine bias. https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing"},{"key":"e_1_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3449148"},{"key":"e_1_2_1_7_1","doi-asserted-by":"crossref","first-page":"171802","DOI":"10.1103\/PhysRevLett.125.171802","article-title":"Sensei: Direct-detection results on sub-gev dark matter from a new skipper ccd","volume":"125","author":"Barak Liron","year":"2020","unstructured":"Liron Barak, Itay M Bloch, Mariano Cababie, Gustavo Cancelo, Luke Chaplinsky, Fernando Chierchie, Michael Crisler, Alex Drlica-Wagner, Rouven Essig, Juan Estrada, et al., 2020. Sensei: Direct-detection results on sub-gev dark matter from a new skipper ccd. Physical Review Letters, Vol. 125, 17 (2020), 171802.","journal-title":"Physical Review Letters"},{"key":"e_1_2_1_8_1","first-page":"1","volume-title":"Proceedings of the ACM on Human-Computer Interaction","volume":"4","author":"Barlas Pinar","year":"2021","unstructured":"Pinar Barlas, Kyriakos Kyriakou, Olivia Guest, Styliani Kleanthous, and Jahna Otterbacher. 2021. To'' see'' is to stereotype: Image tagging algorithms, gender recognition, and the accuracy-fairness trade-off. Proceedings of the ACM on Human-Computer Interaction, Vol. 4, CSCW3 (2021), 1-31."},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.3917\/hori.005.0017"},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1177\/0049124118782533"},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1093\/oxrep\/grab020"},{"key":"e_1_2_1_12_1","volume-title":"Using thematic analysis in psychology. Qualitative research in psychology","author":"Braun Virginia","year":"2006","unstructured":"Virginia Braun and Victoria Clarke. 2006. Using thematic analysis in psychology. Qualitative research in psychology, Vol. 3, 2 (2006), 77-101."},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3236024.3264838"},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.5555\/3272525.3272526"},{"key":"e_1_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3128572.3140448"},{"key":"e_1_2_1_16_1","volume-title":"Fairness testing: A comprehensive survey and analysis of trends. arXiv preprint arXiv:2207.10223","author":"Chen Zhenpeng","year":"2022","unstructured":"Zhenpeng Chen, Jie M Zhang, Max Hort, Federica Sarro, and Mark Harman. 2022. Fairness testing: A comprehensive survey and analysis of trends. arXiv preprint arXiv:2207.10223 (2022)."},{"key":"e_1_2_1_17_1","volume-title":"Congress.","author":"US","year":"2023","unstructured":"US Congress. 2023. H.R.3369 - Artificial intelligence accountability. https:\/\/www.congress. gov\/bill\/118th- congress\/house- bill\/3369\/text"},{"key":"e_1_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3461702.3462519"},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3301275.3302310"},{"volume-title":"Divining a digital future: Mess and mythology in ubiquitous computing","author":"Dourish Paul","key":"e_1_2_1_20_1","unstructured":"Paul Dourish and Genevieve Bell. 2011. Divining a digital future: Mess and mythology in ubiquitous computing. Mit Press."},{"key":"e_1_2_1_21_1","unstructured":"DSS. 1998. LSAC. https:\/\/www.kaggle.com\/datasets\/danofer\/law-school-admissions-bar-passage"},{"key":"e_1_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/2090236.2090255"},{"key":"e_1_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1609\/icwsm.v11i1.14898"},{"key":"e_1_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3510003.3510137"},{"key":"e_1_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/2783258.2783311"},{"key":"e_1_2_1_26_1","volume-title":"Fair AI: Challenges and opportunities. Business & information systems engineering","author":"Feuerriegel Stefan","year":"2020","unstructured":"Stefan Feuerriegel, Mateusz Dolata, and Gerhard Schwabe. 2020. Fair AI: Challenges and opportunities. Business & information systems engineering, Vol. 62 (2020), 379-384."},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3461702.3462621"},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3106237.3106277"},{"key":"e_1_2_1_29_1","doi-asserted-by":"crossref","first-page":"1491","DOI":"10.3390\/diagnostics14141491","article-title":"Evaluating Large Language Model (LLM) Performance on Established Breast Classification Systems","volume":"14","author":"Haider Syed Ali","year":"2024","unstructured":"Syed Ali Haider, Sophia M Pressman, Sahar Borna, Cesar A Gomez-Cabello, Ajai Sehgal, Bradley C Leibovich, and Antonio Jorge Forte. 2024. Evaluating Large Language Model (LLM) Performance on Established Breast Classification Systems. Diagnostics, Vol. 14, 14 (2024), 1491.","journal-title":"Diagnostics"},{"key":"e_1_2_1_30_1","volume-title":"Equality of opportunity in supervised learning. Advances in neural information processing systems","author":"Hardt Moritz","year":"2016","unstructured":"Moritz Hardt, Eric Price, and Nati Srebro. 2016. Equality of opportunity in supervised learning. Advances in neural information processing systems, Vol. 29 (2016)."},{"key":"e_1_2_1_31_1","volume-title":"2015 IEEE 8th International Conference on Software Testing, Verification and Validation (ICST). IEEE, 1-12","author":"Harman Mark","year":"2015","unstructured":"Mark Harman, Yue Jia, and Yuanyuan Zhang. 2015. Achievements, open problems and challenges for search based software testing. In 2015 IEEE 8th International Conference on Software Testing, Verification and Validation (ICST). IEEE, 1-12."},{"key":"e_1_2_1_32_1","volume-title":"Metric learning for individual fairness. arXiv preprint arXiv:1906.00250","author":"Ilvento Christina","year":"2019","unstructured":"Christina Ilvento. 2019. Metric learning for individual fairness. arXiv preprint arXiv:1906.00250 (2019)."},{"key":"e_1_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3579610"},{"key":"e_1_2_1_34_1","volume-title":"Proceedings of the 2nd international workshop on fairness, accountability, transparency and ethics in multimedia. 1-5.","author":"Joo Jungseock","year":"2020","unstructured":"Jungseock Joo and Kimmo K\u00e4rkk\u00e4inen. 2020. Gender slopes: Counterfactual fairness for computer vision models by attribute manipulation. In Proceedings of the 2nd international workshop on fairness, accountability, transparency and ethics in multimedia. 1-5."},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1287\/mnsc.2020.3850"},{"key":"e_1_2_1_36_1","first-page":"857","article-title":"Data-driven discrimination at work","volume":"58","author":"Kim Pauline T","year":"2016","unstructured":"Pauline T Kim. 2016. Data-driven discrimination at work. Wm. & Mary L. Rev., Vol. 58 (2016), 857.","journal-title":"Wm. & Mary L. Rev."},{"key":"e_1_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/2815675.2815714"},{"key":"e_1_2_1_38_1","volume-title":"Proceedings of the 9th Annual Conference on Genetic and Evolutionary Computation. 1098-1105","author":"Lakhotia Kiran","year":"2007","unstructured":"Kiran Lakhotia, Mark Harman, and Phil McMinn. 2007. A multi-objective approach to search-based test data generation. In Proceedings of the 9th Annual Conference on Genetic and Evolutionary Computation. 1098-1105."},{"key":"e_1_2_1_39_1","first-page":"1","volume-title":"Proceedings of the ACM on Human-Computer Interaction","volume":"6","author":"Lam Michelle S","year":"2022","unstructured":"Michelle S Lam, Mitchell L Gordon, Dana\u00eb Metaxa, Jeffrey T Hancock, James A Landay, and Michael S Bernstein. 2022. End-user audits: A system empowering communities to lead large-scale investigations of harmful algorithmic behavior. Proceedings of the ACM on Human-Computer Interaction, Vol. 6, CSCW2 (2022), 1-34."},{"key":"e_1_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1287\/mnsc.2018.3093"},{"key":"e_1_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00459"},{"key":"e_1_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3555150"},{"key":"e_1_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3512899"},{"key":"e_1_2_1_44_1","volume-title":"Wagner Meira Jr, and Virgilio Almeida.","author":"Magno Gabriel","year":"2016","unstructured":"Gabriel Magno, Camila Souza Ara\u00fajo, Wagner Meira Jr, and Virgilio Almeida. 2016. Stereotypes in search engine results: understanding the role of local and global factors. arXiv preprint arXiv:1609.05413 (2016)."},{"key":"e_1_2_1_45_1","unstructured":"Subha Maity Songkai Xue Mikhail Yurochkin and Yuekai Sun. 2020. Statistical inference for individual fairness. (2020)."},{"key":"e_1_2_1_46_1","volume-title":"2022 IEEE World AI IoT Congress. IEEE, 699-707","author":"Mashhadi Afra","year":"2022","unstructured":"Afra Mashhadi, Ali Tabaraei, Yuting Zhan, and Reza M Parizi. 2022. An auditing framework for analyzing fairness of spatial-temporal federated learning applications. In 2022 IEEE World AI IoT Congress. IEEE, 699-707."},{"key":"e_1_2_1_47_1","volume-title":"Prediction Sensitivity: Continual audit of counterfactual fairness in deployed classifiers. arXiv preprint arXiv:2202.04504","author":"Maughan Krystal","year":"2022","unstructured":"Krystal Maughan, Ivoline C Ngong, and Joseph P Near. 2022. Prediction Sensitivity: Continual audit of counterfactual fairness in deployed classifiers. arXiv preprint arXiv:2202.04504 (2022)."},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306618.3317964"},{"key":"e_1_2_1_49_1","volume-title":"Conference on Fairness, accountability and transparency. PMLR, 107-118","author":"Menon Aditya Krishna","year":"2018","unstructured":"Aditya Krishna Menon and Robert C Williamson. 2018. The cost of fairness in binary classification. In Conference on Fairness, accountability and transparency. PMLR, 107-118."},{"key":"e_1_2_1_50_1","first-page":"1","volume-title":"Proceedings of the ACM on Human-Computer Interaction","volume":"5","author":"Metaxa Dana\u00eb","year":"2021","unstructured":"Dana\u00eb Metaxa, Michelle A Gan, Su Goh, Jeff Hancock, and James A Landay. 2021. An image of society: Gender and racial representation and impact in image search results for occupations. Proceedings of the ACM on Human-Computer Interaction, Vol. 5, CSCW1 (2021), 1-23."},{"key":"e_1_2_1_51_1","volume-title":"The New York Times","volume":"6","author":"Metz Cade","year":"2020","unstructured":"Cade Metz and Adam Satariano. 2020. An algorithm that grants freedom, or takes it away. The New York Times, Vol. 6 (2020)."},{"key":"e_1_2_1_52_1","volume-title":"International Conference on Machine Learning. PMLR, 7097-7107","author":"Mukherjee Debarghya","year":"2020","unstructured":"Debarghya Mukherjee, Mikhail Yurochkin, Moulinath Banerjee, and Yuekai Sun. 2020. Two simple ways to learn individual fairness metrics from data. In International Conference on Machine Learning. PMLR, 7097-7107."},{"key":"e_1_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3025453.3025727"},{"key":"e_1_2_1_54_1","unstructured":"European Parliament. 2018. 2018 reform of EU data protection rules. https:\/\/commission.europa.eu\/law\/law-topic\/data-protection\/data-protection-eu_en"},{"key":"e_1_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/1357054.1357160"},{"key":"e_1_2_1_56_1","volume-title":"Blackbox post-processing for multiclass fairness. arXiv preprint arXiv:2201.04461","author":"Putzel Preston","year":"2022","unstructured":"Preston Putzel and Scott Lee. 2022. Blackbox post-processing for multiclass fairness. arXiv preprint arXiv:2201.04461 (2022)."},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3274417"},{"key":"e_1_2_1_58_1","first-page":"15421","volume-title":"Proceedings of the AAAI 37th Conference on Artificial Intelligence","volume":"37","author":"Ruggieri Salvatore","year":"2023","unstructured":"Salvatore Ruggieri, Jose M Alvarez, Andrea Pugnana, Franco Turini, et al., 2023. Can we trust fair-AI?. In Proceedings of the AAAI 37th Conference on Artificial Intelligence, Vol. 37. 15421-15430."},{"key":"e_1_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/3498361.3538788"},{"key":"e_1_2_1_60_1","first-page":"4349","article-title":"Auditing algorithms: Research methods for detecting discrimination on internet platforms","volume":"22","author":"Sandvig Christian","year":"2014","unstructured":"Christian Sandvig, Kevin Hamilton, Karrie Karahalios, and Cedric Langbort. 2014. Auditing algorithms: Research methods for detecting discrimination on internet platforms. Data and Discrimination: Converting Critical Concerns into Productive Inquiry, Vol. 22, 2014 (2014), 4349-4357.","journal-title":"Data and Discrimination: Converting Critical Concerns into Productive Inquiry"},{"key":"e_1_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3479577"},{"key":"e_1_2_1_62_1","volume-title":"The rising safety concerns of deep recommender systems. The Innovation","author":"Shen Huawei","year":"2025","unstructured":"Huawei Shen, Yuanhao Liu, Kaike Zhang, Qi Cao, and Xueqi Cheng. 2025. The rising safety concerns of deep recommender systems. The Innovation (2025), 101038."},{"key":"e_1_2_1_63_1","unstructured":"B Justice Srikrishna. 2018. A free and fair digital economy: Protecting privacy empowering Indians."},{"key":"e_1_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01231-1_31"},{"key":"e_1_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1145\/2447976.2447990"},{"key":"e_1_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445927"},{"key":"e_1_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1145\/3449130"},{"key":"e_1_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300683"},{"key":"e_1_2_1_69_1","unstructured":"UCI. 1994. Statlog (German Credit Data). https:\/\/archive.ics.uci.edu\/dataset\/144\/statloggermancreditdata"},{"key":"e_1_2_1_70_1","unstructured":"UCI. 1996. Adult. https:\/\/archive.ics.uci.edu\/dataset\/2\/adult"},{"key":"e_1_2_1_71_1","unstructured":"UCI. 1998. Estimation of Obesity Levels Based On Eating Habits and Physical Condition. https:\/\/archive.ics.uci.edu\/dataset\/544\/estimationofobesitylevelsbasedoneatinghabitsandphysicalcondition"},{"key":"e_1_2_1_72_1","unstructured":"UCI. 2000. Census-Income (KDD). https:\/\/archive.ics.uci.edu\/dataset\/117\/censusincomekdd"},{"key":"e_1_2_1_73_1","unstructured":"UCI. 2016. Default of credit card clients. https:\/\/archive.ics.uci.edu\/dataset\/350\/defaultofcreditcardclients"},{"key":"e_1_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1145\/3238147.3238165"},{"key":"e_1_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1177\/01655515221093029"},{"key":"e_1_2_1_76_1","unstructured":"Richard Vanderford. 2022. New York's landmark AI bias law prompts uncertainty. https:\/\/www.wsj.com\/articles\/new-yorks-landmark-ai-bias-law-prompts-uncertainty-11663752602"},{"key":"e_1_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445928"},{"key":"e_1_2_1_78_1","volume-title":"International Conference on Artificial Intelligence and Statistics. PMLR, 4552-4562","author":"Xue Songkai","year":"2020","unstructured":"Songkai Xue, Mikhail Yurochkin, and Yuekai Sun. 2020. Auditing ml models for individual bias and unfairness. In International Conference on Artificial Intelligence and Statistics. PMLR, 4552-4562."},{"key":"e_1_2_1_79_1","volume-title":"International Conference on Machine Learning. PMLR, 24929-24962","author":"Yan Tom","year":"2022","unstructured":"Tom Yan and Chicheng Zhang. 2022. Active fairness auditing. In International Conference on Machine Learning. PMLR, 24929-24962."},{"key":"e_1_2_1_80_1","volume-title":"Training individually fair ML models with sensitive subspace robustness. arXiv preprint arXiv:1907.00020","author":"Yurochkin Mikhail","year":"2019","unstructured":"Mikhail Yurochkin, Amanda Bower, and Yuekai Sun. 2019. Training individually fair ML models with sensitive subspace robustness. arXiv preprint arXiv:1907.00020 (2019)."},{"key":"e_1_2_1_81_1","volume-title":"Sensei: Sensitive set invariance for enforcing individual fairness. arXiv preprint arXiv:2006.14168","author":"Yurochkin Mikhail","year":"2020","unstructured":"Mikhail Yurochkin and Yuekai Sun. 2020. Sensei: Sensitive set invariance for enforcing individual fairness. arXiv preprint arXiv:2006.14168 (2020)."},{"key":"e_1_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1145\/3038912.3052660"},{"key":"e_1_2_1_83_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377811.3380331"},{"key":"e_1_2_1_84_1","doi-asserted-by":"crossref","first-page":"3593","DOI":"10.1109\/TSE.2021.3101478","article-title":"Automatic fairness testing of neural classifiers through adversarial sampling","volume":"48","author":"Zhang Peixin","year":"2021","unstructured":"Peixin Zhang, Jingyi Wang, Jun Sun, Xinyu Wang, Guoliang Dong, Xingen Wang, Ting Dai, and Jin Song Dong. 2021. Automatic fairness testing of neural classifiers through adversarial sampling. IEEE Transactions on Software Engineering, Vol. 48, 9 (2021), 3593-3612.","journal-title":"IEEE Transactions on Software Engineering"},{"key":"e_1_2_1_85_1","volume-title":"Pushing The Limit of LLM Capacity for Text Classification. arXiv preprint arXiv:2402.07470","author":"Zhang Yazhou","year":"2024","unstructured":"Yazhou Zhang, Mengyao Wang, Chenyu Ren, Qiuchi Li, Prayag Tiwari, Benyou Wang, and Jing Qin. 2024. Pushing The Limit of LLM Capacity for Text Classification. arXiv preprint arXiv:2402.07470 (2024)."}],"container-title":["Proceedings of the ACM on Human-Computer Interaction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3757414","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,17]],"date-time":"2025-10-17T02:06:35Z","timestamp":1760666795000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3757414"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,16]]},"references-count":85,"journal-issue":{"issue":"7","published-print":{"date-parts":[[2025,10,18]]}},"alternative-id":["10.1145\/3757414"],"URL":"https:\/\/doi.org\/10.1145\/3757414","relation":{},"ISSN":["2573-0142"],"issn-type":[{"type":"electronic","value":"2573-0142"}],"subject":[],"published":{"date-parts":[[2025,10,16]]},"assertion":[{"value":"2025-10-16","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}