{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T20:53:59Z","timestamp":1776113639677,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":78,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T00:00:00Z","timestamp":1686528000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,6,12]]},"DOI":"10.1145\/3593013.3594075","type":"proceedings-article","created":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T14:40:46Z","timestamp":1686580846000},"page":"1213-1228","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":6,"title":["Disentangling and Operationalizing AI Fairness at LinkedIn"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-5149-3228","authenticated-orcid":false,"given":"Joaquin","family":"Qui\u00f1onero Candela","sequence":"first","affiliation":[{"name":"LinkedIn, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-9673-3058","authenticated-orcid":false,"given":"Yuwen","family":"Wu","sequence":"additional","affiliation":[{"name":"LinkedIn, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-0599-3258","authenticated-orcid":false,"given":"Brian","family":"Hsu","sequence":"additional","affiliation":[{"name":"LinkedIn, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7046-9280","authenticated-orcid":false,"given":"Sakshi","family":"Jain","sequence":"additional","affiliation":[{"name":"LinkedIn, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-9656-785X","authenticated-orcid":false,"given":"Jennifer","family":"Ramos","sequence":"additional","affiliation":[{"name":"LinkedIn, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-9682-6561","authenticated-orcid":false,"given":"Jon","family":"Adams","sequence":"additional","affiliation":[{"name":"LinkedIn, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-0124-8689","authenticated-orcid":false,"given":"Robert","family":"Hallman","sequence":"additional","affiliation":[{"name":"LinkedIn, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4091-0119","authenticated-orcid":false,"given":"Kinjal","family":"Basu","sequence":"additional","affiliation":[{"name":"LinkedIn, USA"}]}],"member":"320","published-online":{"date-parts":[[2023,6,12]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"2021. Together we can improve equal access to opportunity. (2021). https:\/\/members.linkedin.com\/equal-access"},{"key":"e_1_3_2_1_2_1","unstructured":"2022. Hide candidate names and photos in LinkedIn Recruiter. (2022). https:\/\/www.linkedin.com\/help\/recruiter\/answer\/a481559"},{"key":"e_1_3_2_1_3_1","unstructured":"2022. Improve Gender Representation in Your Candidate Pools with These Diversity Features. (2022). https:\/\/www.linkedin.com\/pulse\/improve-gender-representation-your-candidate-\/"},{"key":"e_1_3_2_1_4_1","unstructured":"2022. Microsoft Responsible AI Standard. (2022). https:\/\/query.prod.cms.rt.microsoft.com\/cms\/api\/am\/binary\/RE5cmFl"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3219847"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2009.52"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/2623330.2623362"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/2783258.2788614"},{"key":"e_1_3_2_1_9_1","unstructured":"Parag Agrawal Ankan Saha Yafei Wang Aastha Nigam and Eric Lawrence. 2020. Building a heterogeneous social network recommendation system. (2020). https:\/\/engineering.linkedin.com\/blog\/2020\/building-a-heterogeneous-social-network-recommendation-system"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSEC.2018.2888775"},{"key":"e_1_3_2_1_11_1","volume-title":"Can explainable AI explain unfairness? A framework for evaluating explainable AI. arXiv preprint arXiv:2106.07483","author":"Alikhademi Kiana","year":"2021","unstructured":"Kiana Alikhademi, Brianna Richardson, Emma Drobina, and Juan E Gilbert. 2021. Can explainable AI explain unfairness? A framework for evaluating explainable AI. arXiv preprint arXiv:2106.07483 (2021)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445888"},{"key":"e_1_3_2_1_13_1","volume-title":"Joaquin Qui\u00f1onero Candela","author":"Bakalar Chlo\u00e9","year":"2021","unstructured":"Chlo\u00e9 Bakalar, Renata Barreto, Stevie Bergman, Miranda Bogen, Bobbie Chern, Sam Corbett-Davies, Melissa Hall, Isabel Kloumann, Michelle Lam, Joaquin Qui\u00f1onero Candela, 2021. Fairness on the ground: Applying algorithmic fairness approaches to production systems. arXiv preprint arXiv:2103.06172 (2021)."},{"key":"e_1_3_2_1_14_1","volume-title":"Fairness in machine learning. Nips tutorial 1","author":"Barocas Solon","year":"2017","unstructured":"Solon Barocas, Moritz Hardt, and Arvind Narayanan. 2017. Fairness in machine learning. Nips tutorial 1 (2017), 2."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"crossref","unstructured":"Antonio Bella C\u00e8sar Ferri Jos\u00e9 Hern\u00e1ndez-Orallo and Mar\u00eda Jos\u00e9 Ram\u00edrez-Quintana. 2010. Calibration of machine learning models. In Handbook of Research on Machine Learning Applications and Trends: Algorithms Methods and Techniques. IGI Global 128\u2013146.","DOI":"10.4018\/978-1-60566-766-9.ch006"},{"key":"e_1_3_2_1_16_1","volume-title":"Chi","author":"Beutel Alex","year":"2019","unstructured":"Alex Beutel, Jilin Chen, Tulsee Doshi, Hai Qian, Allison Woodruff, Christine Luu, Pierre Kreitmann, Jonathan Bischof, and Ed H. Chi. 2019. Putting Fairness Principles into Practice: Challenges, Metrics, and Improvements. https:\/\/arxiv.org\/pdf\/1901.04562.pdf"},{"key":"e_1_3_2_1_17_1","volume-title":"Mehdi Mashayekhi, Qi Zeng, Rabyd (Rob) Roudani, Sean Gahagan, Andrew Howard, and Isabella Leone.","author":"Bogen Miranda","year":"2023","unstructured":"Miranda Bogen, Pushkar Tripathi, Aditya Srinivas Timmaraju, Mehdi Mashayekhi, Qi Zeng, Rabyd (Rob) Roudani, Sean Gahagan, Andrew Howard, and Isabella Leone. 2023. Toward fairness in personalized ads. (2023). https:\/\/about.fb.com\/news\/2023\/01\/an-update-on-our-ads-fairness-efforts\/"},{"key":"e_1_3_2_1_18_1","volume-title":"Inference for individual mediation effects and interventional effects in sparse high-dimensional causal graphical models. arXiv preprint arXiv:1809.10652","author":"Chakrabortty Abhishek","year":"2018","unstructured":"Abhishek Chakrabortty, Preetam Nandy, and Hongzhe Li. 2018. Inference for individual mediation effects and interventional effects in sparse high-dimensional causal graphical models. arXiv preprint arXiv:1809.10652 (2018)."},{"key":"e_1_3_2_1_19_1","volume-title":"Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2","author":"Chouldechova Alexandra","year":"2017","unstructured":"Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2 (2017), 153\u2013163."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3376898"},{"key":"e_1_3_2_1_21_1","volume-title":"The measure and mismeasure of fairness: A critical review of fair machine learning. arXiv preprint arXiv:1808.00023","author":"Corbett-Davies Sam","year":"2018","unstructured":"Sam Corbett-Davies and Sharad Goel. 2018. The measure and mismeasure of fairness: A critical review of fair machine learning. arXiv preprint arXiv:1808.00023 (2018)."},{"key":"e_1_3_2_1_22_1","volume-title":"The atlas of AI: Power, politics, and the planetary costs of artificial intelligence","author":"Crawford Kate","unstructured":"Kate Crawford. 2021. The atlas of AI: Power, politics, and the planetary costs of artificial intelligence. Yale University Press."},{"key":"e_1_3_2_1_23_1","volume-title":"Predictive Rate Parity Testing and Mitigation. arXiv preprint arXiv:2204.05947","author":"DiCiccio Cyrus","year":"2022","unstructured":"Cyrus DiCiccio, Brian Hsu, YinYin Yu, Preetam Nandy, and Kinjal Basu. 2022. Predictive Rate Parity Testing and Mitigation. arXiv preprint arXiv:2204.05947 (2022)."},{"key":"e_1_3_2_1_24_1","unstructured":"Imani Dunbar. 2022. Mythbusting the Feed: How We Work to Address Bias. (2022). https:\/\/blog.linkedin.com\/2022\/november\/1\/mythbusting-the-feed-how-we-work-to-address-bias"},{"key":"e_1_3_2_1_25_1","unstructured":"Imani Dunbar. 2022. Scaling Self-ID to Better Connect Members to Opportunities. (2022). https:\/\/www.linkedin.com\/pulse\/scaling-self-id-better-connect-members-opportunities-imani-dunbar\/?trk=pulse-article"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-79228-4_1"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/2090236.2090255"},{"key":"e_1_3_2_1_28_1","volume-title":"The algorithmic foundations of differential privacy. Foundations and Trends\u00ae in Theoretical Computer Science 9, 3\u20134","author":"Dwork Cynthia","year":"2014","unstructured":"Cynthia Dwork, Aaron Roth, 2014. The algorithmic foundations of differential privacy. Foundations and Trends\u00ae in Theoretical Computer Science 9, 3\u20134 (2014), 211\u2013407."},{"key":"e_1_3_2_1_29_1","volume-title":"Pascal Van Hentenryck, and Keyu Zhu","author":"Fioretto Ferdinando","year":"2022","unstructured":"Ferdinando Fioretto, Cuong Tran, Pascal Van Hentenryck, and Keyu Zhu. 2022. Differential Privacy and Fairness in Decisions and Learning Tasks: A Survey. arXiv preprint arXiv:2202.08187 (2022)."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.5555\/2907333.2907524"},{"key":"e_1_3_2_1_31_1","volume-title":"Representation-Aware Experimentation: Group Inequality Analysis for A\/B Testing and Alerting. arXiv preprint arXiv:2204.12011","author":"Friedberg Rina","year":"2022","unstructured":"Rina Friedberg, Stuart Ambler, and Guillaume Saint-Jacques. 2022. Representation-Aware Experimentation: Group Inequality Analysis for A\/B Testing and Alerting. arXiv preprint arXiv:2204.12011 (2022)."},{"key":"e_1_3_2_1_32_1","volume-title":"Privacy Aware Experimentation over Sensitive Groups: A General Chi Square Approach. arXiv preprint arXiv:2208.08564","author":"Friedberg Rina","year":"2022","unstructured":"Rina Friedberg and Ryan Rogers. 2022. Privacy Aware Experimentation over Sensitive Groups: A General Chi Square Approach. arXiv preprint arXiv:2208.08564 (2022)."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306618.3317950"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3210205"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3514094.3534157"},{"key":"e_1_3_2_1_36_1","first-page":"1","article-title":"A survey of learning causality with data: Problems and methods","volume":"53","author":"Guo Ruocheng","year":"2020","unstructured":"Ruocheng Guo, Lu Cheng, Jundong Li, P Richard Hahn, and Huan Liu. 2020. A survey of learning causality with data: Problems and methods. ACM Computing Surveys (CSUR) 53, 4 (2020), 1\u201337.","journal-title":"ACM Computing Surveys (CSUR)"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/2959100.2959196"},{"key":"e_1_3_2_1_38_1","volume-title":"Balancing out bias: Achieving fairness through training reweighting. arXiv preprint arXiv:2109.08253","author":"Han Xudong","year":"2021","unstructured":"Xudong Han, Timothy Baldwin, and Trevor Cohn. 2021. Balancing out bias: Achieving fairness through training reweighting. arXiv preprint arXiv:2109.08253 (2021)."},{"key":"e_1_3_2_1_39_1","volume-title":"Equality of opportunity in supervised learning. Advances in neural information processing systems 29","author":"Hardt Moritz","year":"2016","unstructured":"Moritz Hardt, Eric Price, and Nati Srebro. 2016. Equality of opportunity in supervised learning. Advances in neural information processing systems 29 (2016)."},{"key":"e_1_3_2_1_40_1","volume-title":"International Conference on Machine Learning. PMLR","author":"Hashimoto Tatsunori","year":"2018","unstructured":"Tatsunori Hashimoto, Megha Srivastava, Hongseok Namkoong, and Percy Liang. 2018. Fairness without demographics in repeated loss minimization. In International Conference on Machine Learning. PMLR, 1929\u20131938."},{"key":"e_1_3_2_1_41_1","volume-title":"International Conference on Machine Learning. PMLR","author":"H\u00e9bert-Johnson Ursula","year":"2018","unstructured":"Ursula H\u00e9bert-Johnson, Michael Kim, Omer Reingold, and Guy Rothblum. 2018. Multicalibration: Calibration for the (computationally-identifiable) masses. In International Conference on Machine Learning. PMLR, 1939\u20131948."},{"key":"e_1_3_2_1_42_1","unstructured":"Sara Hooker Dumitru Erhan Pieter-Jan Kindermans and Been Kim. 2018. Evaluating feature importance estimates. (2018)."},{"key":"e_1_3_2_1_43_1","volume-title":"An Operational Perspective to Fairness Interventions: Where and How to Intervene. arXiv preprint arXiv:2302.01574","author":"Hsu Brian","year":"2023","unstructured":"Brian Hsu, Xiaotong Chen, Ying Han, Hongseok Namkoong, and Kinjal Basu. 2023. An Operational Perspective to Fairness Interventions: Where and How to Intervene. arXiv preprint arXiv:2302.01574 (2023)."},{"key":"e_1_3_2_1_44_1","volume-title":"Pushing the limits of fairness impossibility: Who\u2019s the fairest of them all?arXiv preprint arXiv:2208.12606","author":"Hsu Brian","year":"2022","unstructured":"Brian Hsu, Rahul Mazumder, Preetam Nandy, and Kinjal Basu. 2022. Pushing the limits of fairness impossibility: Who\u2019s the fairest of them all?arXiv preprint arXiv:2208.12606 (2022)."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357974"},{"key":"e_1_3_2_1_46_1","volume-title":"International conference on machine learning. PMLR, 1617\u20131626","author":"Jabbari Shahin","year":"2017","unstructured":"Shahin Jabbari, Matthew Joseph, Michael Kearns, Jamie Morgenstern, and Aaron Roth. 2017. Fairness in reinforcement learning. In International conference on machine learning. PMLR, 1617\u20131626."},{"key":"e_1_3_2_1_47_1","volume-title":"International Conference on Machine Learning. PMLR, 4721\u20134732","author":"Jalal Ajil","year":"2021","unstructured":"Ajil Jalal, Sushrut Karmalkar, Jessica Hoffmann, Alex Dimakis, and Eric Price. 2021. Fairness for image generation with uncertain sensitive attributes. In International Conference on Machine Learning. PMLR, 4721\u20134732."},{"key":"e_1_3_2_1_48_1","volume-title":"Fairness in learning: Classic and contextual bandits. Advances in neural information processing systems 29","author":"Joseph Matthew","year":"2016","unstructured":"Matthew Joseph, Michael Kearns, Jamie H Morgenstern, and Aaron Roth. 2016. Fairness in learning: Classic and contextual bandits. Advances in neural information processing systems 29 (2016)."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/3109859.3109921"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306618.3314287"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.4230\/LIPIcs.ITCS.2017.43"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/2487575.2488217"},{"key":"e_1_3_2_1_53_1","volume-title":"Chi","author":"Lahoti Preethi","year":"2020","unstructured":"Preethi Lahoti, Alex Beutel, Jilin Chen, Kang Lee, Flavien Prost, Nithum Thain, Xuezhi Wang, and Ed Chi. 2020. Fairness without demographics through adversarially reweighted learning. Advances in neural information processing systems 33 (2020), 728\u2013740."},{"key":"e_1_3_2_1_54_1","volume-title":"Explore-Exploit: A Framework for Interactive and Online Learning. arXiv preprint arXiv:1812.00116","author":"Liu Honglei","year":"2018","unstructured":"Honglei Liu, Anuj Kumar, Wenhai Yang, and Benoit Dumoulin. 2018. Explore-Exploit: A Framework for Interactive and Online Learning. arXiv preprint arXiv:1812.00116 (2018)."},{"key":"e_1_3_2_1_55_1","volume-title":"Manish Bhide, Diptikalyan Saha, Kush R Varshney, and Ruchir Puri.","author":"Lohia Pranay K","year":"2019","unstructured":"Pranay K Lohia, Karthikeyan Natesan Ramamurthy, Manish Bhide, Diptikalyan Saha, Kush R Varshney, and Ruchir Puri. 2019. Bias mitigation post-processing for individual and group fairness. In Icassp 2019-2019 ieee international conference on acoustics, speech and signal processing (icassp). IEEE, 2847\u20132851."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2017.12"},{"key":"e_1_3_2_1_57_1","first-page":"2870","article-title":"A\/B testing in dense large-scale networks: design and inference","volume":"33","author":"Nandy Preetam","year":"2020","unstructured":"Preetam Nandy, Kinjal Basu, Shaunak Chatterjee, and Ye Tu. 2020. A\/B testing in dense large-scale networks: design and inference. Advances in Neural Information Processing Systems 33 (2020), 2870\u20132880.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_58_1","unstructured":"Preetam Nandy Yunsong Meng Cyrus DiCiccio Heloise Logan Amir Sepehri Divya Venugopalan Kinjal Basu and Noureddine El Karoui. 2021. Using the LinkedIn Fairness Toolkit in large-scale AI systems. (2021). https:\/\/engineering.linkedin.com\/blog\/2021\/using-the-linkedin-fairness-toolkit-large-scale-ai"},{"key":"e_1_3_2_1_59_1","first-page":"6466","article-title":"A\/B testing for recommender systems in a two-sided marketplace","volume":"34","author":"Nandy Preetam","year":"2021","unstructured":"Preetam Nandy, Divya Venugopalan, Chun Lo, and Shaunak Chatterjee. 2021. A\/B testing for recommender systems in a two-sided marketplace. Advances in Neural Information Processing Systems 34 (2021), 6466\u20136477.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_60_1","unstructured":"Arvind Narayanan. 21. Fairness definitions and their politics. In Tutorial presented at the Conf. on Fairness Accountability and Transparency."},{"key":"e_1_3_2_1_61_1","unstructured":"Judea Pearl. 2009. Causality. Cambridge university press."},{"key":"e_1_3_2_1_62_1","first-page":"25944","article-title":"Post-processing for individual fairness","volume":"34","author":"Petersen Felix","year":"2021","unstructured":"Felix Petersen, Debarghya Mukherjee, Yuekai Sun, and Mikhail Yurochkin. 2021. Post-processing for individual fairness. Advances in Neural Information Processing Systems 34 (2021), 25944\u201325955.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_63_1","volume-title":"On fairness and calibration. Advances in neural information processing systems 30","author":"Pleiss Geoff","year":"2017","unstructured":"Geoff Pleiss, Manish Raghavan, Felix Wu, Jon Kleinberg, and Kilian Q Weinberger. 2017. On fairness and calibration. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372828"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1145\/3269206.3272030"},{"key":"e_1_3_2_1_66_1","volume-title":"An Outcome Test of Discrimination for Ranked Lists. In 2022 ACM Conference on Fairness, Accountability, and Transparency. 350\u2013356","author":"Roth Jonathan","year":"2022","unstructured":"Jonathan Roth, Guillaume Saint-Jacques, and YinYin Yu. 2022. An Outcome Test of Discrimination for Ranked Lists. In 2022 ACM Conference on Fairness, Accountability, and Transparency. 350\u2013356."},{"key":"e_1_3_2_1_67_1","volume-title":"Fairness through Experimentation: Inequality in A\/B testing as an approach to responsible design. arXiv preprint arXiv:2002.05819","author":"Saint-Jacques Guillaume","year":"2020","unstructured":"Guillaume Saint-Jacques, Amir Sepehri, Nicole Li, and Igor Perisic. 2020. Fairness through Experimentation: Inequality in A\/B testing as an approach to responsible design. arXiv preprint arXiv:2002.05819 (2020)."},{"key":"e_1_3_2_1_68_1","volume-title":"Secretary Problems with Biased Evaluations using Partial Ordinal Information. Available at SSRN 3444283","author":"Salem Jad","year":"2019","unstructured":"Jad Salem and Swati Gupta. 2019. Secretary Problems with Biased Evaluations using Partial Ordinal Information. Available at SSRN 3444283 (2019)."},{"key":"e_1_3_2_1_69_1","volume-title":"Prompting gpt-3 to be reliable. arXiv preprint arXiv:2210.09150","author":"Si Chenglei","year":"2022","unstructured":"Chenglei Si, Zhe Gan, Zhengyuan Yang, Shuohang Wang, Jianfeng Wang, Jordan Boyd-Graber, and Lijuan Wang. 2022. Prompting gpt-3 to be reliable. arXiv preprint arXiv:2210.09150 (2022)."},{"key":"e_1_3_2_1_70_1","first-page":"352","article-title":"Private machine learning classification based on fully homomorphic encryption","volume":"8","author":"Sun Xiaoqiang","year":"2018","unstructured":"Xiaoqiang Sun, Peng Zhang, Joseph K Liu, Jianping Yu, and Weixin Xie. 2018. Private machine learning classification based on fully homomorphic encryption. IEEE Transactions on Emerging Topics in Computing 8, 2 (2018), 352\u2013364.","journal-title":"IEEE Transactions on Emerging Topics in Computing"},{"key":"e_1_3_2_1_71_1","volume-title":"Improving the fairness of deep generative models without retraining. arXiv preprint arXiv:2012.04842","author":"Tan Shuhan","year":"2020","unstructured":"Shuhan Tan, Yujun Shen, and Bolei Zhou. 2020. Improving the fairness of deep generative models without retraining. arXiv preprint arXiv:2012.04842 (2020)."},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3412705"},{"key":"e_1_3_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.clsr.2021.105567"},{"key":"e_1_3_2_1_74_1","volume-title":"In-Processing Modeling Techniques for Machine Learning Fairness: A Survey. ACM Transactions on Knowledge Discovery from Data (TKDD)","author":"Wan Mingyang","year":"2022","unstructured":"Mingyang Wan, Daochen Zha, Ninghao Liu, and Na Zou. 2022. In-Processing Modeling Techniques for Machine Learning Fairness: A Survey. ACM Transactions on Knowledge Discovery from Data (TKDD) (2022)."},{"key":"e_1_3_2_1_75_1","volume-title":"On the legal compatibility of fairness definitions. arXiv preprint arXiv:1912.00761","author":"Xiang Alice","year":"2019","unstructured":"Alice Xiang and Inioluwa Deborah Raji. 2019. On the legal compatibility of fairness definitions. arXiv preprint arXiv:1912.00761 (2019)."},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-32236-6_51"},{"key":"e_1_3_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDCS.2015.40"},{"key":"e_1_3_2_1_78_1","volume-title":"Homomorphic encryption","author":"Yi Xun","unstructured":"Xun Yi, Russell Paulet, Elisa Bertino, Xun Yi, Russell Paulet, and Elisa Bertino. 2014. Homomorphic encryption. Springer."}],"event":{"name":"FAccT '23: the 2023 ACM Conference on Fairness, Accountability, and Transparency","location":"Chicago IL USA","acronym":"FAccT '23"},"container-title":["2023 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594075","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3593013.3594075","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:18Z","timestamp":1750178238000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594075"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,12]]},"references-count":78,"alternative-id":["10.1145\/3593013.3594075","10.1145\/3593013"],"URL":"https:\/\/doi.org\/10.1145\/3593013.3594075","relation":{},"subject":[],"published":{"date-parts":[[2023,6,12]]},"assertion":[{"value":"2023-06-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}