{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,28]],"date-time":"2025-08-28T12:14:39Z","timestamp":1756383279144,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":38,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T00:00:00Z","timestamp":1730073600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Key R&D Program of China","award":["2023YFC3310700"],"award-info":[{"award-number":["2023YFC3310700"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,28]]},"DOI":"10.1145\/3664647.3681524","type":"proceedings-article","created":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T06:59:27Z","timestamp":1729925967000},"page":"1866-1874","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Poisoning for Debiasing: Fair Recognition via Eliminating Bias Uncovered in Data Poisoning"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6586-1037","authenticated-orcid":false,"given":"Yi","family":"Zhang","sequence":"first","affiliation":[{"name":"Huawei Cloud, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6703-2064","authenticated-orcid":false,"given":"Zhefeng","family":"Wang","sequence":"additional","affiliation":[{"name":"Huawei Cloud, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-7891-1950","authenticated-orcid":false,"given":"Rui","family":"Hu","sequence":"additional","affiliation":[{"name":"Beijing Jiaotong University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6803-7964","authenticated-orcid":false,"given":"Xinyu","family":"Duan","sequence":"additional","affiliation":[{"name":"Huawei Cloud, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3890-7575","authenticated-orcid":false,"given":"Yi","family":"Zheng","sequence":"additional","affiliation":[{"name":"Huawei Cloud, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9625-2314","authenticated-orcid":false,"given":"Baoxing","family":"Huai","sequence":"additional","affiliation":[{"name":"Huawei Cloud, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-1006-8719","authenticated-orcid":false,"given":"Jiarun","family":"Han","sequence":"additional","affiliation":[{"name":"Beijing Jiaotong University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0699-3205","authenticated-orcid":false,"given":"Jitao","family":"Sang","sequence":"additional","affiliation":[{"name":"Beijing Jiaotong University &amp; Peng Cheng Lab, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2024,10,28]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"International Conference on Machine Learning. PMLR, 528--539","author":"Bahng Hyojin","year":"2020","unstructured":"Hyojin Bahng, Sanghyuk Chun, Sangdoo Yun, Jaegul Choo, and Seong Joon Oh. 2020. Learning de-biased representations with biased representations. In International Conference on Machine Learning. PMLR, 528--539."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3585385"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2008.11"},{"key":"e_1_3_2_1_4_1","volume-title":"The accuracy, fairness, and limits of predicting recidivism. Science advances","author":"Dressel Julia","year":"2018","unstructured":"Julia Dressel and Hany Farid. 2018. The accuracy, fairness, and limits of predicting recidivism. Science advances, Vol. 4, 1 (2018), eaao5580."},{"key":"e_1_3_2_1_5_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=SJx0q1rtvS","author":"Du Min","year":"2020","unstructured":"Min Du, Ruoxi Jia, and Dawn Song. 2020. Robust anomaly detection and backdoor attack detection via differential privacy. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=SJx0q1rtvS"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359789.3359790"},{"key":"e_1_3_2_1_7_1","volume-title":"Badnets: Identifying vulnerabilities in the machine learning model supply chain. arXiv preprint arXiv:1708.06733","author":"Gu Tianyu","year":"2017","unstructured":"Tianyu Gu, Brendan Dolan-Gavitt, and Siddharth Garg. 2017. Badnets: Identifying vulnerabilities in the machine learning model supply chain. arXiv preprint arXiv:1708.06733 (2017)."},{"key":"e_1_3_2_1_8_1","volume-title":"Equality of opportunity in supervised learning. Advances in neural information processing systems","author":"Hardt Moritz","year":"2016","unstructured":"Moritz Hardt, Eric Price, and Nati Srebro. 2016. Equality of opportunity in supervised learning. Advances in neural information processing systems, Vol. 29 (2016)."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_10_1","first-page":"26449","article-title":"Unbiased classification through bias-contrastive and bias-balanced learning","volume":"34","author":"Hong Youngkyu","year":"2021","unstructured":"Youngkyu Hong and Eunho Yang. 2021. Unbiased classification through bias-contrastive and bias-balanced learning. Advances in Neural Information Processing Systems, Vol. 34 (2021), 26449--26461.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3581783.3612312"},{"key":"e_1_3_2_1_12_1","volume-title":"James Bailey, and Yisen Wang.","author":"Huang Hanxun","year":"2021","unstructured":"Hanxun Huang, Xingjun Ma, Sarah Monazam Erfani, James Bailey, and Yisen Wang. 2021. Unlearnable Examples: Making Personal Data Unexploitable. In ICLR."},{"key":"e_1_3_2_1_13_1","volume-title":"Advances in Neural Information Processing Systems","volume":"36","author":"Jha Rishi","year":"2024","unstructured":"Rishi Jha, Jonathan Hayase, and Sewoong Oh. 2024. Label poisoning is all you need. Advances in Neural Information Processing Systems, Vol. 36 (2024)."},{"key":"e_1_3_2_1_14_1","volume-title":"International Conference on Machine Learning. PMLR, 15435--15450","author":"Jung Yeonsung","year":"2023","unstructured":"Yeonsung Jung, Hajin Shim, June Yong Yang, and Eunho Yang. 2023. Fighting fire with fire: contrastive debiasing without bias-free data via generative bias-transformation. In International Conference on Machine Learning. PMLR, 15435--15450."},{"key":"e_1_3_2_1_15_1","volume-title":"Supervised contrastive learning. Advances in neural information processing systems","author":"Khosla Prannay","year":"2020","unstructured":"Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. 2020. Supervised contrastive learning. Advances in neural information processing systems, Vol. 33 (2020), 18661--18673."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00922"},{"key":"e_1_3_2_1_17_1","first-page":"25123","article-title":"Learning debiased representation via disentangled feature augmentation","volume":"34","author":"Lee Jungsoo","year":"2021","unstructured":"Jungsoo Lee, Eungyeup Kim, Juyoung Lee, Jihyeon Lee, and Jaegul Choo. 2021. Learning debiased representation via disentangled feature augmentation. Advances in Neural Information Processing Systems, Vol. 34 (2021), 25123--25133.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i12.26748"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19778-9_16"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.324"},{"key":"e_1_3_2_1_21_1","volume-title":"International Conference on Machine Learning. PMLR, 6781--6792","author":"Liu Evan Z","year":"2021","unstructured":"Evan Z Liu, Behzad Haghgoo, Annie S Chen, Aditi Raghunathan, Pang Wei Koh, Shiori Sagawa, Percy Liang, and Chelsea Finn. 2021. Just train twice: Improving group robustness without training group information. In International Conference on Machine Learning. PMLR, 6781--6792."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.425"},{"key":"e_1_3_2_1_23_1","first-page":"20673","article-title":"Learning from failure: De-biasing classifier from biased classifier","volume":"33","author":"Nam Junhyun","year":"2020","unstructured":"Junhyun Nam, Hyuntak Cha, Sungsoo Ahn, Jaeho Lee, and Jinwoo Shin. 2020. Learning from failure: De-biasing classifier from biased classifier. Advances in Neural Information Processing Systems, Vol. 33 (2020), 20673--20684.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01014"},{"key":"e_1_3_2_1_25_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=ryxGuJrFvS","author":"Shiori","year":"2020","unstructured":"Shiori Sagawa*, Pang Wei Koh*, Tatsunori B. Hashimoto, and Percy Liang. 2020. Distributionally robust neural networks for group shifts: On the importance of regularization for worst-case generalization. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=ryxGuJrFvS"},{"key":"e_1_3_2_1_26_1","volume-title":"Learning from others' mistakes: Avoiding dataset biases without modeling them. arXiv preprint arXiv:2012.01300","author":"Sanh Victor","year":"2020","unstructured":"Victor Sanh, Thomas Wolf, Yonatan Belinkov, and Alexander M Rush. 2020. Learning from others' mistakes: Avoiding dataset biases without modeling them. arXiv preprint arXiv:2012.01300 (2020)."},{"volume-title":"Unsupervised Learning of Debiased Representations with Pseudo-Attributes. In 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society","author":"Seo S.","key":"e_1_3_2_1_27_1","unstructured":"S. Seo, J. Lee, and B. Han. 2022. Unsupervised Learning of Debiased Representations with Pseudo-Attributes. In 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, Los Alamitos, CA, USA, 16721--16730."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i2.20115"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01624"},{"key":"e_1_3_2_1_30_1","volume-title":"Poison frogs! targeted clean-label poisoning attacks on neural networks. Advances in neural information processing systems","author":"Shafahi Ali","year":"2018","unstructured":"Ali Shafahi, W Ronny Huang, Mahyar Najibi, Octavian Suciu, Christoph Studer, Tudor Dumitras, and Tom Goldstein. 2018. Poison frogs! targeted clean-label poisoning attacks on neural networks. Advances in neural information processing systems, Vol. 31 (2018)."},{"key":"e_1_3_2_1_31_1","volume-title":"USENIX Sec. Symp. USENIX Association","volume":"8","author":"Shan Shawn","year":"2022","unstructured":"Shawn Shan, Arjun Nitin Bhagoji, Haitao Zheng, and Ben Y Zhao. 2022. Traceback of targeted data poisoning attacks in neural networks. In USENIX Sec. Symp. USENIX Association, Vol. 8."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01330"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3103191"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00541"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3278721.3278779"},{"key":"e_1_3_2_1_36_1","volume-title":"Inference-Time Rule Eraser: Fair Recognition via Distilling and Removing Biased Rules. arXiv preprint arXiv:2404.04814","author":"Zhang Yi","year":"2024","unstructured":"Yi Zhang, Dongyuan Lu, and Jitao Sang. 2024. Inference-Time Rule Eraser: Fair Recognition via Distilling and Removing Biased Rules. arXiv preprint arXiv:2404.04814 (2024)."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3581783.3612317"},{"key":"e_1_3_2_1_38_1","volume-title":"Generalized cross entropy loss for training deep neural networks with noisy labels. Advances in neural information processing systems","author":"Zhang Zhilu","year":"2018","unstructured":"Zhilu Zhang and Mert Sabuncu. 2018. Generalized cross entropy loss for training deep neural networks with noisy labels. Advances in neural information processing systems, Vol. 31 (2018)."}],"event":{"name":"MM '24: The 32nd ACM International Conference on Multimedia","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Melbourne VIC Australia","acronym":"MM '24"},"container-title":["Proceedings of the 32nd ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681524","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3664647.3681524","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:57:48Z","timestamp":1750294668000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681524"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,28]]},"references-count":38,"alternative-id":["10.1145\/3664647.3681524","10.1145\/3664647"],"URL":"https:\/\/doi.org\/10.1145\/3664647.3681524","relation":{},"subject":[],"published":{"date-parts":[[2024,10,28]]},"assertion":[{"value":"2024-10-28","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}