{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:14:33Z","timestamp":1750220073140,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":78,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,7]],"date-time":"2022-06-07T00:00:00Z","timestamp":1654560000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,7]]},"DOI":"10.1145\/3532105.3535023","type":"proceedings-article","created":{"date-parts":[[2022,6,8]],"date-time":"2022-06-08T14:29:57Z","timestamp":1654698597000},"page":"173-184","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["FairRoad: Achieving Fairness for Recommender Systems with Optimized Antidote Data"],"prefix":"10.1145","author":[{"given":"Minghong","family":"Fang","sequence":"first","affiliation":[{"name":"The Ohio State University, Columbus, OH, USA"}]},{"given":"Jia","family":"Liu","sequence":"additional","affiliation":[{"name":"The Ohio State University, Columbus, OH, USA"}]},{"given":"Michinari","family":"Momma","sequence":"additional","affiliation":[{"name":"Amazon, Seattle, WA, USA"}]},{"given":"Yi","family":"Sun","sequence":"additional","affiliation":[{"name":"Amazon, Seattle, WA, USA"}]}],"member":"320","published-online":{"date-parts":[[2022,6,8]]},"reference":[{"doi-asserted-by":"publisher","key":"e_1_3_2_2_1_1","DOI":"10.5555\/3016100.3016102"},{"key":"e_1_3_2_2_2_1","volume-title":"Discrimination through optimization: How Facebook's ad delivery can lead to skewed outcomes. arXiv preprint arXiv:1904.02095 (April","author":"Ali Muhammad","year":"2019","unstructured":"Muhammad Ali , Piotr Sapiezynski , Miranda Bogen , Aleksandra Korolova , Alan Mislove , and Aaron Rieke . 2019. Discrimination through optimization: How Facebook's ad delivery can lead to skewed outcomes. arXiv preprint arXiv:1904.02095 (April 2019 ). Muhammad Ali, Piotr Sapiezynski, Miranda Bogen, Aleksandra Korolova, Alan Mislove, and Aaron Rieke. 2019. Discrimination through optimization: How Facebook's ad delivery can lead to skewed outcomes. arXiv preprint arXiv:1904.02095 (April 2019)."},{"unstructured":"Gender API. [Online]. https:\/\/gender-api.com  Gender API. [Online]. https:\/\/gender-api.com","key":"e_1_3_2_2_3_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_4_1","DOI":"10.1177\/0049124118782533"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_5_1","DOI":"10.1145\/3292500.3330745"},{"unstructured":"Battista Biggio Blaine Nelson and Pavel Laskov. 2012. Poisoning attacks against support vector machines. In ICML.  Battista Biggio Blaine Nelson and Pavel Laskov. 2012. Poisoning attacks against support vector machines. In ICML.","key":"e_1_3_2_2_6_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_7_1","DOI":"10.1145\/3351095.3372864"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_8_1","DOI":"10.1145\/3368089.3409704"},{"key":"e_1_3_2_2_9_1","volume-title":"Individually Fair Ranking. arXiv preprint arXiv:2103.11023","author":"Bower Amanda","year":"2021","unstructured":"Amanda Bower , Hamid Eftekhari , Mikhail Yurochkin , and Yuekai Sun . 2021. Individually Fair Ranking. arXiv preprint arXiv:2103.11023 ( 2021 ). Amanda Bower, Hamid Eftekhari, Mikhail Yurochkin, and Yuekai Sun. 2021. Individually Fair Ranking. arXiv preprint arXiv:2103.11023 (2021)."},{"key":"e_1_3_2_2_10_1","volume-title":"Conference on Fairness, Accountability and Transparency. PMLR, 202--214","author":"Burke Robin","year":"2018","unstructured":"Robin Burke , Nasim Sonboli , and Aldo Ordonez-Gauger . 2018 . Balanced neighborhoods for multi-sided fairness in recommendation . In Conference on Fairness, Accountability and Transparency. PMLR, 202--214 . Robin Burke, Nasim Sonboli, and Aldo Ordonez-Gauger. 2018. Balanced neighborhoods for multi-sided fairness in recommendation. In Conference on Fairness, Accountability and Transparency. PMLR, 202--214."},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_11_1","DOI":"10.1109\/ICDMW.2009.83"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_12_1","DOI":"10.1007\/s10618-010-0190-x"},{"unstructured":"Xiaoyu Cao Minghong Fang Jia Liu and Neil Zhenqiang Gong. 2021. FLTrust: Byzantine-robust Federated Learning via Trust Bootstrapping. In NDSS.  Xiaoyu Cao Minghong Fang Jia Liu and Neil Zhenqiang Gong. 2021. FLTrust: Byzantine-robust Federated Learning via Trust Bootstrapping. In NDSS.","key":"e_1_3_2_2_13_1"},{"key":"e_1_3_2_2_14_1","volume-title":"30th USENIX Security Symposium (USENIX Security 21)","author":"Cao Xiaoyu","year":"2021","unstructured":"Xiaoyu Cao , Jinyuan Jia , and Neil Zhenqiang Gong . 2021 . Data poisoning attacks to local differential privacy protocols . In 30th USENIX Security Symposium (USENIX Security 21) . 947--964. Xiaoyu Cao, Jinyuan Jia, and Neil Zhenqiang Gong. 2021. Data poisoning attacks to local differential privacy protocols. In 30th USENIX Security Symposium (USENIX Security 21). 947--964."},{"key":"e_1_3_2_2_15_1","volume-title":"Bias and debias in recommender system: A survey and future directions. arXiv preprint arXiv:2010.03240","author":"Chen Jiawei","year":"2020","unstructured":"Jiawei Chen , Hande Dong , Xiang Wang , Fuli Feng , Meng Wang , and Xiangnan He. 2020. Bias and debias in recommender system: A survey and future directions. arXiv preprint arXiv:2010.03240 ( 2020 ). Jiawei Chen, Hande Dong, Xiang Wang, Fuli Feng, Meng Wang, and Xiangnan He. 2020. Bias and debias in recommender system: A survey and future directions. arXiv preprint arXiv:2010.03240 (2020)."},{"key":"e_1_3_2_2_16_1","volume-title":"Targeted backdoor attacks on deep learning systems using data poisoning. arXiv preprint arXiv:1712.05526","author":"Chen Xinyun","year":"2017","unstructured":"Xinyun Chen , Chang Liu , Bo Li , Kimberly Lu , and Dawn Song . 2017. Targeted backdoor attacks on deep learning systems using data poisoning. arXiv preprint arXiv:1712.05526 ( 2017 ). Xinyun Chen, Chang Liu, Bo Li, Kimberly Lu, and Dawn Song. 2017. Targeted backdoor attacks on deep learning systems using data poisoning. arXiv preprint arXiv:1712.05526 (2017)."},{"unstructured":"Jaewoong Cho Moonseok Choi and Changho Suh. 2021. Equal Experience in Recommender Systems. (2021).  Jaewoong Cho Moonseok Choi and Changho Suh. 2021. Equal Experience in Recommender Systems. (2021).","key":"e_1_3_2_2_17_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_18_1","DOI":"10.1145\/2090236.2090255"},{"key":"e_1_3_2_2_19_1","volume-title":"29th USENIX Security Symposium (USENIX Security 20)","author":"Fang Minghong","year":"2020","unstructured":"Minghong Fang , Xiaoyu Cao , Jinyuan Jia , and Neil Gong . 2020 . Local Model Poisoning Attacks to Byzantine-Robust Federated Learning . In 29th USENIX Security Symposium (USENIX Security 20) . 1605--1622. Minghong Fang, Xiaoyu Cao, Jinyuan Jia, and Neil Gong. 2020. Local Model Poisoning Attacks to Byzantine-Robust Federated Learning. In 29th USENIX Security Symposium (USENIX Security 20). 1605--1622."},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_20_1","DOI":"10.1145\/3366423.3380072"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_21_1","DOI":"10.1145\/3442381.3450066"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_22_1","DOI":"10.1145\/3274694.3274706"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_23_1","DOI":"10.1145\/3397271.3401051"},{"key":"e_1_3_2_2_24_1","volume-title":"Counteracting Bias and Increasing Fairness in Search and Recommender Systems. In Fourteenth ACM Conference on Recommender Systems. 745--747","author":"Gao Ruoyuan","year":"2020","unstructured":"Ruoyuan Gao and Chirag Shah . 2020 . Counteracting Bias and Increasing Fairness in Search and Recommender Systems. In Fourteenth ACM Conference on Recommender Systems. 745--747 . Ruoyuan Gao and Chirag Shah. 2020. Counteracting Bias and Increasing Fairness in Search and Recommender Systems. In Fourteenth ACM Conference on Recommender Systems. 745--747."},{"doi-asserted-by":"crossref","unstructured":"David Garcia-Soriano and Francesco Bonchi. 2021. Maxmin-Fair Ranking: Individual Fairness under Group-Fairness Constraints. In KDD.  David Garcia-Soriano and Francesco Bonchi. 2021. Maxmin-Fair Ranking: Individual Fairness under Group-Fairness Constraints. In KDD.","key":"e_1_3_2_2_25_1","DOI":"10.1145\/3447548.3467349"},{"unstructured":"Moritz Hardt Eric Price and Nathan Srebro. 2016. Equality of opportunity in supervised learning. In NeurIPS.  Moritz Hardt Eric Price and Nathan Srebro. 2016. Equality of opportunity in supervised learning. In NeurIPS.","key":"e_1_3_2_2_26_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_27_1","DOI":"10.1145\/3442381.3449904"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_28_1","DOI":"10.1109\/SP.2018.00057"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_29_1","DOI":"10.1145\/3460120.3485368"},{"key":"e_1_3_2_2_30_1","volume-title":"Proceedings of the AAAI Conference on Artificial Intelligence","volume":"35","author":"Jang Taeuk","year":"2021","unstructured":"Taeuk Jang , Feng Zheng , and Xiaoqian Wang . 2021 . Constructing a Fair Classifier with Generated Fair Data . In Proceedings of the AAAI Conference on Artificial Intelligence , Vol. 35 . 7908--7916. Taeuk Jang, Feng Zheng, and Xiaoqian Wang. 2021. Constructing a Fair Classifier with Generated Fair Data. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35. 7908--7916."},{"key":"e_1_3_2_2_31_1","volume-title":"Ananya Kumar, and Percy Liang.","author":"Jones Erik","year":"2020","unstructured":"Erik Jones , Shiori Sagawa , Pang Wei Koh , Ananya Kumar, and Percy Liang. 2020 . Selective classification can magnify disparities across groups. arXiv preprint arXiv:2010.14134 (2020). Erik Jones, Shiori Sagawa, Pang Wei Koh, Ananya Kumar, and Percy Liang. 2020. Selective classification can magnify disparities across groups. arXiv preprint arXiv:2010.14134 (2020)."},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_32_1","DOI":"10.1007\/s10115-011-0463-8"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_33_1","DOI":"10.1145\/3394486.3403080"},{"key":"e_1_3_2_2_34_1","volume-title":"International Conference on Machine Learning. PMLR, 5264--5274","author":"Kim Joon Sik","year":"2020","unstructured":"Joon Sik Kim , Jiahao Chen , and Ameet Talwalkar . 2020 . Fact: A diagnostic for group fairness trade-offs . In International Conference on Machine Learning. PMLR, 5264--5274 . Joon Sik Kim, Jiahao Chen, and Ameet Talwalkar. 2020. Fact: A diagnostic for group fairness trade-offs. In International Conference on Machine Learning. PMLR, 5264--5274."},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_35_1","DOI":"10.1145\/3442381.3450080"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_36_1","DOI":"10.1007\/s10994-021-06119-y"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_37_1","DOI":"10.1109\/MC.2009.263"},{"unstructured":"Bo Li Yining Wang Aarti Singh and Yevgeniy Vorobeychik. 2016. Data poisoning attacks on factorization-based collaborative filtering. In Advances in neural information processing systems. 1885--1893.  Bo Li Yining Wang Aarti Singh and Yevgeniy Vorobeychik. 2016. Data poisoning attacks on factorization-based collaborative filtering. In Advances in neural information processing systems. 1885--1893.","key":"e_1_3_2_2_38_1"},{"unstructured":"Tian Li Maziar Sanjabi Ahmad Beirami and Virginia Smith. 2020. Fair resource allocation in federated learning. In ICLR.  Tian Li Maziar Sanjabi Ahmad Beirami and Virginia Smith. 2020. Fair resource allocation in federated learning. In ICLR.","key":"e_1_3_2_2_39_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_40_1","DOI":"10.1145\/3366424.3383557"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_41_1","DOI":"10.1145\/3394486.3403102"},{"key":"e_1_3_2_2_42_1","volume-title":"Bias disparity in collaborative recommendation: Algorithmic evaluation and comparison. arXiv preprint arXiv:1908.00831","author":"Mansoury Masoud","year":"2019","unstructured":"Masoud Mansoury , Bamshad Mobasher , Robin Burke , and Mykola Pechenizkiy . 2019. Bias disparity in collaborative recommendation: Algorithmic evaluation and comparison. arXiv preprint arXiv:1908.00831 ( 2019 ). Masoud Mansoury, Bamshad Mobasher, Robin Burke, and Mykola Pechenizkiy. 2019. Bias disparity in collaborative recommendation: Algorithmic evaluation and comparison. arXiv preprint arXiv:1908.00831 (2019)."},{"key":"e_1_3_2_2_43_1","volume-title":"Thomas","author":"Metevier Blossom","year":"2019","unstructured":"Blossom Metevier , Stephen Giguere , Sarah Brockman , Ari Kobren , Yuriy Brun , Emma Brunskill , and Philip S . Thomas . 2019 . Offline Contextual Bandits with High Probability Fairness Guarantees. In Proc. NeurIPS. Blossom Metevier, Stephen Giguere, Sarah Brockman, Ari Kobren, Yuriy Brun, Emma Brunskill, and Philip S. Thomas. 2019. Offline Contextual Bandits with High Probability Fairness Guarantees. In Proc. NeurIPS."},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_44_1","DOI":"10.1007\/s00146-020-00950-y"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_45_1","DOI":"10.1145\/3397271.3401100"},{"key":"e_1_3_2_2_46_1","volume-title":"Achieving fairness via post-processing in web-scale recommender systems. arXiv preprint arXiv:2006.11350","author":"Nandy Preetam","year":"2020","unstructured":"Preetam Nandy , Cyrus Diciccio , Divya Venugopalan , Heloise Logan , Kinjal Basu , and Noureddine El Karoui . 2020. Achieving fairness via post-processing in web-scale recommender systems. arXiv preprint arXiv:2006.11350 ( 2020 ). Preetam Nandy, Cyrus Diciccio, Divya Venugopalan, Heloise Logan, Kinjal Basu, and Noureddine El Karoui. 2020. Achieving fairness via post-processing in web-scale recommender systems. arXiv preprint arXiv:2006.11350 (2020)."},{"key":"e_1_3_2_2_47_1","volume-title":"Study: Facebook Delivers Biased Job Ads, Skewed by Gender. https:\/\/apnews.com\/article\/discrimination-f62160cbbad4d72ce5250e6ef2222f5e","author":"O'Brien Matt","year":"2021","unstructured":"Matt O'Brien and Barbara Ortutay . 2021 . Study: Facebook Delivers Biased Job Ads, Skewed by Gender. https:\/\/apnews.com\/article\/discrimination-f62160cbbad4d72ce5250e6ef2222f5e Matt O'Brien and Barbara Ortutay. 2021. Study: Facebook Delivers Biased Job Ads, Skewed by Gender. https:\/\/apnews.com\/article\/discrimination-f62160cbbad4d72ce5250e6ef2222f5e"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_48_1","DOI":"10.1145\/3366423.3380196"},{"key":"e_1_3_2_2_49_1","volume-title":"Post-processing for Individual Fairness. Advances in Neural Information Processing Systems 34","author":"Petersen Felix","year":"2021","unstructured":"Felix Petersen , Debarghya Mukherjee , Yuekai Sun , and Mikhail Yurochkin . 2021. Post-processing for Individual Fairness. Advances in Neural Information Processing Systems 34 ( 2021 ). Felix Petersen, Debarghya Mukherjee, Yuekai Sun, and Mikhail Yurochkin. 2021. Post-processing for Individual Fairness. Advances in Neural Information Processing Systems 34 (2021)."},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_50_1","DOI":"10.1145\/3289600.3291002"},{"key":"e_1_3_2_2_51_1","volume-title":"International Conference on Machine Learning. PMLR, 9389--9398","author":"Schwarzschild Avi","year":"2021","unstructured":"Avi Schwarzschild , Micah Goldblum , Arjun Gupta , John P Dickerson , and Tom Goldstein . 2021 . Just how toxic is data poisoning? a unified benchmark for backdoor and data poisoning attacks . In International Conference on Machine Learning. PMLR, 9389--9398 . Avi Schwarzschild, Micah Goldblum, Arjun Gupta, John P Dickerson, and Tom Goldstein. 2021. Just how toxic is data poisoning? a unified benchmark for backdoor and data poisoning attacks. In International Conference on Machine Learning. PMLR, 9389--9398."},{"unstructured":"Ali Shafahi W Ronny Huang Mahyar Najibi Octavian Suciu Christoph Studer Tudor Dumitras and Tom Goldstein. 2018. Poison frogs! targeted clean-label poisoning attacks on neural networks. In Advances in Neural Information Processing Systems. 6103--6113.  Ali Shafahi W Ronny Huang Mahyar Najibi Octavian Suciu Christoph Studer Tudor Dumitras and Tom Goldstein. 2018. Poison frogs! targeted clean-label poisoning attacks on neural networks. In Advances in Neural Information Processing Systems. 6103--6113.","key":"e_1_3_2_2_52_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_53_1","DOI":"10.1145\/3461702.3462559"},{"key":"e_1_3_2_2_54_1","volume-title":"Poisoning attacks on algorithmic fairness. arXiv preprint arXiv:2004.07401","author":"Solans David","year":"2020","unstructured":"David Solans , Battista Biggio , and Carlos Castillo . 2020. Poisoning attacks on algorithmic fairness. arXiv preprint arXiv:2004.07401 ( 2020 ). David Solans, Battista Biggio, and Carlos Castillo. 2020. Poisoning attacks on algorithmic fairness. arXiv preprint arXiv:2004.07401 (2020)."},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_55_1","DOI":"10.1145\/3219819.3220046"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_56_1","DOI":"10.1145\/3240323.3240372"},{"key":"e_1_3_2_2_57_1","article-title":"Visualizing data using t-SNE","volume":"9","author":"der Maaten Laurens Van","year":"2008","unstructured":"Laurens Van der Maaten and Geoffrey Hinton . 2008 . Visualizing data using t-SNE . Journal of machine learning research 9 , 11 (2008). Laurens Van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE. Journal of machine learning research 9, 11 (2008).","journal-title":"Journal of machine learning research"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_58_1","DOI":"10.1145\/3336191.3371855"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_59_1","DOI":"10.1145\/3442188.3445915"},{"doi-asserted-by":"crossref","unstructured":"Wenjie Wang Fuli Feng Xiangnan He Xiang Wang and Tat-Seng Chua. 2021. Deconfounded Recommendation for Alleviating Bias Amplification. In KDD.  Wenjie Wang Fuli Feng Xiangnan He Xiang Wang and Tat-Seng Chua. 2021. Deconfounded Recommendation for Alleviating Bias Amplification. In KDD.","key":"e_1_3_2_2_60_1","DOI":"10.1145\/3447548.3467249"},{"key":"e_1_3_2_2_61_1","volume-title":"Data poisoning attacks against online learning. arXiv preprint arXiv:1808.08994","author":"Wang Yizhen","year":"2018","unstructured":"Yizhen Wang and Kamalika Chaudhuri . 2018. Data poisoning attacks against online learning. arXiv preprint arXiv:1808.08994 ( 2018 ). Yizhen Wang and Kamalika Chaudhuri. 2018. Data poisoning attacks against online learning. arXiv preprint arXiv:1808.08994 (2018)."},{"unstructured":"Zirui Wang Yulia Tsvetkov Orhan Firat and Yuan Cao. 2021. Gradient vaccine: Investigating and improving multi-task optimization in massively multilingual models. In ICLR.  Zirui Wang Yulia Tsvetkov Orhan Firat and Yuan Cao. 2021. Gradient vaccine: Investigating and improving multi-task optimization in massively multilingual models. In ICLR.","key":"e_1_3_2_2_62_1"},{"unstructured":"Kyle Wiggers. 2020. Researchers Find Evidence of Bias in Recommender Systems. https:\/\/venturebeat.com\/2020\/07\/29\/researchers-find-evidence-of-bias-in-recommender-systems\/  Kyle Wiggers. 2020. Researchers Find Evidence of Bias in Recommender Systems. https:\/\/venturebeat.com\/2020\/07\/29\/researchers-find-evidence-of-bias-in-recommender-systems\/","key":"e_1_3_2_2_63_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_64_1","DOI":"10.1145\/3447548.3467335"},{"unstructured":"Chuhan Wu Fangzhao Wu Xiting Wang Yongfeng Huang and Xing Xie. 2021. Fairrec: fairness-aware news recommendation with decomposed adversarial learning. In AAAI.  Chuhan Wu Fangzhao Wu Xiting Wang Yongfeng Huang and Xing Xie. 2021. Fairrec: fairness-aware news recommendation with decomposed adversarial learning. In AAAI.","key":"e_1_3_2_2_65_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_66_1","DOI":"10.1145\/3442381.3450015"},{"key":"e_1_3_2_2_67_1","volume-title":"International Conference on Machine Learning. 1689--1698","author":"Xiao Huang","year":"2015","unstructured":"Huang Xiao , Battista Biggio , Gavin Brown , Giorgio Fumera , Claudia Eckert , and Fabio Roli . 2015 . Is feature selection secure against training data poisoning? . In International Conference on Machine Learning. 1689--1698 . Huang Xiao, Battista Biggio, Gavin Brown, Giorgio Fumera, Claudia Eckert, and Fabio Roli. 2015. Is feature selection secure against training data poisoning?. In International Conference on Machine Learning. 1689--1698."},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_68_1","DOI":"10.1145\/3109859.3109887"},{"unstructured":"Sirui Yao and Bert Huang. 2017. Beyond parity: Fairness objectives for collaborative filtering. In NeurIPS.  Sirui Yao and Bert Huang. 2017. Beyond parity: Fairness objectives for collaborative filtering. In NeurIPS.","key":"e_1_3_2_2_69_1"},{"unstructured":"Yelp Challenge Dataset. 2018. https:\/\/www.yelp.com\/dataset\/challenge  Yelp Challenge Dataset. 2018. https:\/\/www.yelp.com\/dataset\/challenge","key":"e_1_3_2_2_70_1"},{"unstructured":"Tianhe Yu Saurabh Kumar Abhishek Gupta Sergey Levine Karol Hausman and Chelsea Finn. 2020. Gradient surgery for multi-task learning. In NeurIPS.  Tianhe Yu Saurabh Kumar Abhishek Gupta Sergey Levine Karol Hausman and Chelsea Finn. 2020. Gradient surgery for multi-task learning. In NeurIPS.","key":"e_1_3_2_2_71_1"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_72_1","DOI":"10.1145\/3038912.3052660"},{"key":"e_1_3_2_2_73_1","volume-title":"From parity to preference-based notions of fairness in classification. Advances in Neural Information Processing Systems 30","author":"Zafar Muhammad Bilal","year":"2017","unstructured":"Muhammad Bilal Zafar , Isabel Valera , Manuel Rodriguez , Krishna Gummadi , and Adrian Weller . 2017. From parity to preference-based notions of fairness in classification. Advances in Neural Information Processing Systems 30 ( 2017 ). Muhammad Bilal Zafar, Isabel Valera, Manuel Rodriguez, Krishna Gummadi, and Adrian Weller. 2017. From parity to preference-based notions of fairness in classification. Advances in Neural Information Processing Systems 30 (2017)."},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_74_1","DOI":"10.1145\/3132847.3132938"},{"key":"e_1_3_2_2_75_1","volume-title":"International conference on machine learning. PMLR, 325--333","author":"Zemel Rich","year":"2013","unstructured":"Rich Zemel , Yu Wu , Kevin Swersky , Toni Pitassi , and Cynthia Dwork . 2013 . Learning fair representations . In International conference on machine learning. PMLR, 325--333 . Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. 2013. Learning fair representations. In International conference on machine learning. PMLR, 325--333."},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_76_1","DOI":"10.1145\/3448016.3452787"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_77_1","DOI":"10.1145\/3437963.3441820"},{"doi-asserted-by":"publisher","key":"e_1_3_2_2_78_1","DOI":"10.1145\/3397271.3401177"}],"event":{"sponsor":["SIGSAC ACM Special Interest Group on Security, Audit, and Control"],"acronym":"SACMAT '22","name":"SACMAT '22: The 27th ACM Symposium on Access Control Models and Technologies","location":"New York NY USA"},"container-title":["Proceedings of the 27th ACM on Symposium on Access Control Models and Technologies"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3532105.3535023","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3532105.3535023","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T18:09:14Z","timestamp":1750183754000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3532105.3535023"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,7]]},"references-count":78,"alternative-id":["10.1145\/3532105.3535023","10.1145\/3532105"],"URL":"https:\/\/doi.org\/10.1145\/3532105.3535023","relation":{},"subject":[],"published":{"date-parts":[[2022,6,7]]},"assertion":[{"value":"2022-06-08","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}