{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,6]],"date-time":"2025-10-06T19:20:28Z","timestamp":1759778428605,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":42,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,1,4]],"date-time":"2024-01-04T00:00:00Z","timestamp":1704326400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,1,4]]},"DOI":"10.1145\/3632410.3633294","type":"proceedings-article","created":{"date-parts":[[2024,1,3]],"date-time":"2024-01-03T18:15:16Z","timestamp":1704305716000},"page":"510-513","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Tutorial on Fair and Private Deep Learning"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8918-4739","authenticated-orcid":false,"given":"Manisha","family":"Padala","sequence":"first","affiliation":[{"name":"Indian Institute of Science, Bangalore, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1460-6102","authenticated-orcid":false,"given":"Sankarshan","family":"Damle","sequence":"additional","affiliation":[{"name":"IIIT Hyderabad, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4634-7862","authenticated-orcid":false,"given":"Sujit","family":"Gujar","sequence":"additional","affiliation":[{"name":"International Institute of Information Technology, Hyderabad, India"}]}],"member":"320","published-online":{"date-parts":[[2024,1,4]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"crossref","unstructured":"Martin Abadi Andy Chu Ian Goodfellow H\u00a0Brendan McMahan Ilya Mironov Kunal Talwar and Li Zhang. 2016. Deep learning with differential privacy. In ACM SIGSAC CCS. 308\u2013318.","DOI":"10.1145\/2976749.2978318"},{"key":"e_1_3_2_1_2_1","unstructured":"Alekh Agarwal Alina Beygelzimer Miroslav Dudik John Langford and Hanna Wallach. 2018. A Reductions Approach to Fair Classification. In ICML. 60\u201369."},{"key":"e_1_3_2_1_3_1","volume-title":"Differential privacy has disparate impact on model accuracy. NeurIPS 32","author":"Bagdasaryan Eugene","year":"2019","unstructured":"Eugene Bagdasaryan, Omid Poursaeed, and Vitaly Shmatikov. 2019. Differential privacy has disparate impact on model accuracy. NeurIPS 32 (2019)."},{"key":"e_1_3_2_1_4_1","first-page":"671","article-title":"Big data\u2019s disparate impact","volume":"104","author":"Barocas Solon","year":"2016","unstructured":"Solon Barocas and Andrew\u00a0D Selbst. 2016. Big data\u2019s disparate impact. Cal. L. Rev. 104 (2016), 671.","journal-title":"Cal. L. Rev."},{"key":"e_1_3_2_1_5_1","volume-title":"Learning Fair Classifiers: A Regularization-Inspired Approach. CoRR abs\/1707.00044","author":"Bechavod Yahav","year":"2017","unstructured":"Yahav Bechavod and Katrina Ligett. 2017. Learning Fair Classifiers: A Regularization-Inspired Approach. CoRR abs\/1707.00044 (2017)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1177\/0049124118782533"},{"key":"e_1_3_2_1_7_1","volume-title":"Data Decisions and Theoretical Implications when Adversarially Learning Fair Representations. CoRR abs\/1707.00075","author":"Beutel Alex","year":"2017","unstructured":"Alex Beutel, Jilin Chen, Zhe Zhao, and Ed\u00a0Huai hsin Chi. 2017. Data Decisions and Theoretical Implications when Adversarially Learning Fair Representations. CoRR abs\/1707.00075 (2017)."},{"key":"e_1_3_2_1_8_1","volume-title":"Fairness Constraints: Mechanisms for Fair Classification. ArXiv e-prints (July","author":"Zafar M. Bilal","year":"2015","unstructured":"M. Bilal Zafar, I. Valera, M. Gomez Rodriguez, and K.\u00a0P. Gummadi. 2015. Fairness Constraints: Mechanisms for Fair Classification. ArXiv e-prints (July 2015). arxiv:1507.05259\u00a0[stat.ML]"},{"key":"e_1_3_2_1_9_1","volume-title":"Fairness in machine learning: A survey. arXiv preprint arXiv:2010.04053","author":"Caton Simon","year":"2020","unstructured":"Simon Caton and Christian Haas. 2020. Fairness in machine learning: A survey. arXiv preprint arXiv:2010.04053 (2020)."},{"key":"e_1_3_2_1_10_1","volume-title":"Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2","author":"Chouldechova Alexandra","year":"2017","unstructured":"Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2 (2017), 153\u2013163."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"crossref","unstructured":"Rachel Cummings Varun Gupta Dhamma Kimpara and Jamie Morgenstern. 2019. On the compatibility of privacy and fairness. In UMAP. 309\u2013315.","DOI":"10.1145\/3314183.3323847"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"crossref","unstructured":"Sankarshan Damle Aleksei Triastcyn Boi Faltings and Sujit Gujar. 2021. Differentially Private Multi-Agent Constraint Optimization. In WI-IAT. 422\u2013429.","DOI":"10.1145\/3486622.3493929"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"crossref","unstructured":"Cynthia Dwork Moritz Hardt Toniann Pitassi Omer Reingold and Richard Zemel. 2012. Fairness through awareness. In ITCS. 214\u2013226.","DOI":"10.1145\/2090236.2090255"},{"key":"e_1_3_2_1_14_1","unstructured":"Harrison Edwards and Amos Storkey. 2016. Censoring Representations with an Adversary. In ICLR. arxiv.org\/abs\/1511.05897"},{"key":"e_1_3_2_1_15_1","volume-title":"NeurIPS Workshop on New Frontiers in Federated Learning (NFFL).","author":"Ezzeldin H","year":"2021","unstructured":"Yahya\u00a0H Ezzeldin, Shen Yan, Chaoyang He, Emilio Ferrara, and Salman Avestimehr. 2021. Fairfed: Enabling group fairness in federated learning. In NeurIPS Workshop on New Frontiers in Federated Learning (NFFL)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"crossref","unstructured":"Michael Feldman Sorelle\u00a0A. Friedler John Moeller Carlos Scheidegger and Suresh Venkatasubramanian. 2015. Certifying and Removing Disparate Impact. In KDD. 259\u2013268.","DOI":"10.1145\/2783258.2783311"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","unstructured":"Ferdinando Fioretto Cuong Tran Pascal Van\u00a0Hentenryck and Keyu Zhu. 2022. Differential Privacy and Fairness in Decisions and Learning Tasks: A Survey. In IJCAI. 5470\u20135477.","DOI":"10.24963\/ijcai.2022\/766"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/2810103.2813677"},{"key":"e_1_3_2_1_19_1","volume-title":"ICLR Workshop on Socially Responsible Machine Learning.","author":"Hu Shengyuan","year":"2022","unstructured":"Shengyuan Hu, Zhiwei\u00a0Steven Wu, and Virginia Smith. 2022. Provably Fair Federated Learning via Bounded Group Loss. In ICLR Workshop on Socially Responsible Machine Learning."},{"volume-title":"2009 2nd International Conference on Computer, Control and Communication. 1\u20136.","author":"Kamiran F.","key":"e_1_3_2_1_20_1","unstructured":"F. Kamiran and T. Calders. 2009. Classifying without discriminating. In 2009 2nd International Conference on Computer, Control and Communication. 1\u20136."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"crossref","unstructured":"Samhita Kanaparthy Manisha Padala Sankarshan Damle and Sujit Gujar. 2022. Fair federated learning for heterogeneous data. In CODS-COMAD. 298\u2013299.","DOI":"10.1145\/3493700.3493750"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"crossref","unstructured":"Samhita Kanaparthy Manisha Padala Sankarshan Damle Ravi\u00a0Kiran Sarvadevabhatla and Sujit Gujar. 2023. F3: fair and federated face attribute classification with heterogeneous data. In PAKDD. 483\u2013494.","DOI":"10.1007\/978-3-031-33374-3_38"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306618.3314287"},{"key":"e_1_3_2_1_24_1","volume-title":"Priority-based post-processing bias mitigation for individual and group fairness. arXiv preprint arXiv:2102.00417","author":"Lohia Pranay","year":"2021","unstructured":"Pranay Lohia. 2021. Priority-based post-processing bias mitigation for individual and group fairness. arXiv preprint arXiv:2102.00417 (2021)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"crossref","unstructured":"Pranay\u00a0K Lohia Karthikeyan\u00a0Natesan Ramamurthy Manish Bhide Diptikalyan Saha Kush\u00a0R Varshney and Ruchir Puri. 2019. Bias mitigation post-processing for individual and group fairness. In ICASSP. 2847\u20132851.","DOI":"10.1109\/ICASSP.2019.8682620"},{"key":"e_1_3_2_1_26_1","unstructured":"David Madras Elliot Creager Toniann Pitassi and Richard\u00a0S. Zemel. 2018. Learning Adversarially Fair and Transferable Representations. In ICML. 3384\u20133393."},{"key":"e_1_3_2_1_27_1","volume-title":"Privacy in deep learning: A survey. arXiv preprint arXiv:2004.12254","author":"Mireshghallah Fatemehsadat","year":"2020","unstructured":"Fatemehsadat Mireshghallah, Mohammadkazem Taram, Praneeth Vepakomma, Abhishek Singh, Ramesh Raskar, and Hadi Esmaeilzadeh. 2020. Privacy in deep learning: A survey. arXiv preprint arXiv:2004.12254 (2020)."},{"key":"e_1_3_2_1_28_1","unstructured":"Hussein Mozannar Mesrob Ohannessian and Nathan Srebro. 2020. Fair learning with private demographic data. In ICML. 7066\u20137075."},{"key":"e_1_3_2_1_29_1","unstructured":"Harikrishna Narasimhan. 2018. Learning with Complex Loss Functions and Constraints. In AISTATS. 1646\u20131654."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"crossref","unstructured":"Manisha Padala Sankarshan Damle and Sujit Gujar. 2021. Federated Learning Meets Fairness and Differential Privacy. In ICONIP. 692\u2013699.","DOI":"10.1007\/978-3-030-92310-5_80"},{"key":"e_1_3_2_1_31_1","volume-title":"FNNC: Achieving Fairness through Neural Networks. In IJCAI. 2277\u20132283.","author":"Padala Manisha","year":"2020","unstructured":"Manisha Padala and Sujit Gujar. 2020. FNNC: Achieving Fairness through Neural Networks. In IJCAI. 2277\u20132283."},{"key":"e_1_3_2_1_32_1","unstructured":"Nicolas Papernot Mart\u00edn Abadi Ulfar Erlingsson Ian Goodfellow and Kunal Talwar. 2017. Semi-supervised Knowledge Transfer for Deep Learning from Private Training Data. In ICLR. openreview.net\/forum?id=HkwoSDPgg"},{"key":"e_1_3_2_1_33_1","unstructured":"Geoff Pleiss Manish Raghavan Felix Wu Jon Kleinberg and Kilian\u00a0Q Weinberger. 2017. On Fairness and Calibration. In Advances in Neural Information Processing Systems 30. 5680\u20135689."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"crossref","unstructured":"Kritika Prakash Fiza Husain Praveen Paruchuri and Sujit Gujar. 2022. How private is your RL policy? An inverse RL based analysis framework. In AAAI Vol.\u00a036. 8009\u20138016.","DOI":"10.1609\/aaai.v36i7.20772"},{"key":"e_1_3_2_1_35_1","volume-title":"FAIR-FATE: Fair Federated Learning with Momentum. In International Conference on Computational Science. 524\u2013538","author":"Salazar Teresa","year":"2023","unstructured":"Teresa Salazar, Miguel Fernandes, Helder Ara\u00fajo, and Pedro\u00a0Henriques Abreu. 2023. FAIR-FATE: Fair Federated Learning with Momentum. In International Conference on Computational Science. 524\u2013538."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"crossref","unstructured":"Sambhav Solanki Samhita Kanaparthy Sankarshan Damle and Sujit Gujar. 2022. Differentially Private Federated Combinatorial Bandits with Constraints. In ECML PKDD. 620\u2013637.","DOI":"10.1007\/978-3-031-26412-2_38"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"crossref","unstructured":"Cuong Tran Ferdinando Fioretto and Pascal Van\u00a0Hentenryck. 2021. Differentially Private and Fair Deep Learning: A Lagrangian Dual Approach. In AAAI. 9932\u20139939.","DOI":"10.1609\/aaai.v35i11.17193"},{"volume-title":"Federated learning with bayesian differential privacy","author":"Triastcyn Aleksei","key":"e_1_3_2_1_38_1","unstructured":"Aleksei Triastcyn and Boi Faltings. 2019. Federated learning with bayesian differential privacy. In IEEE Big Data. 2587\u20132596."},{"key":"e_1_3_2_1_39_1","volume-title":"Optimized score transformation for fair classification. PMLR 108","author":"Wei Dennis","year":"2020","unstructured":"Dennis Wei, Karthikeyan\u00a0Natesan Ramamurthy, and Flavio\u00a0P Calmon. 2020. Optimized score transformation for fair classification. PMLR 108 (2020)."},{"key":"e_1_3_2_1_40_1","volume-title":"Fairness-aware Classification: Criterion, Convexity, and Bounds. CoRR abs\/1809.04737","author":"Wu Yongkai","year":"2018","unstructured":"Yongkai Wu, Lu Zhang, and Xintao Wu. 2018. Fairness-aware Classification: Criterion, Convexity, and Bounds. CoRR abs\/1809.04737 (2018)."},{"key":"e_1_3_2_1_41_1","volume-title":"Mitigating Unwanted Biases with Adversarial Learning. CoRR abs\/1801.07593","author":"Zhang Brian\u00a0Hu","year":"2018","unstructured":"Brian\u00a0Hu Zhang, Blake Lemoine, and Margaret Mitchell. 2018. Mitigating Unwanted Biases with Adversarial Learning. CoRR abs\/1801.07593 (2018). arxiv:1801.07593http:\/\/arxiv.org\/abs\/1801.07593"},{"key":"e_1_3_2_1_42_1","volume-title":"Fairfl: A fair federated learning approach to reducing demographic bias in privacy-sensitive classification models","author":"Zhang Daniel\u00a0Yue","year":"2020","unstructured":"Daniel\u00a0Yue Zhang, Ziyi Kou, and Dong Wang. 2020. Fairfl: A fair federated learning approach to reducing demographic bias in privacy-sensitive classification models. In IEEE Big Data. 1051\u20131060."}],"event":{"name":"CODS-COMAD 2024: 7th Joint International Conference on Data Science & Management of Data (11th ACM IKDD CODS and 29th COMAD)","acronym":"CODS-COMAD 2024","location":"Bangalore India"},"container-title":["Proceedings of the 7th Joint International Conference on Data Science &amp; Management of Data (11th ACM IKDD CODS and 29th COMAD)"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3632410.3633294","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3632410.3633294","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T13:34:52Z","timestamp":1755869692000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3632410.3633294"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,4]]},"references-count":42,"alternative-id":["10.1145\/3632410.3633294","10.1145\/3632410"],"URL":"https:\/\/doi.org\/10.1145\/3632410.3633294","relation":{},"subject":[],"published":{"date-parts":[[2024,1,4]]},"assertion":[{"value":"2024-01-04","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}