{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T03:00:20Z","timestamp":1775617220302,"version":"3.50.1"},"reference-count":83,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,1,3]],"date-time":"2024-01-03T00:00:00Z","timestamp":1704240000000},"content-version":"vor","delay-in-days":2,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100000781","name":"European Research Council","doi-asserted-by":"publisher","award":["741278"],"award-info":[{"award-number":["741278"]}],"id":[{"id":"10.13039\/501100000781","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Empir Software Eng"],"published-print":{"date-parts":[[2024,1]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Decision-making software mainly based on Machine Learning (ML) may contain fairness issues (e.g., providing favourable treatment to certain people rather than others based on sensitive attributes such as gender or race). Various mitigation methods have been proposed to automatically repair fairness issues to achieve fairer ML software and help software engineers to create responsible software. However, existing bias mitigation methods trade accuracy for fairness (i.e., trade a reduction in accuracy for better fairness). In this paper, we present a novel search-based method for repairing ML-based decision making software to simultaneously increase both its fairness and accuracy. As far as we know, this is the first bias mitigation approach based on multi-objective search that aims to repair fairness issues without trading accuracy for binary classification methods. We apply our approach to two widely studied ML models in the software fairness literature (i.e., Logistic Regression and Decision Trees), and compare it with seven publicly available state-of-the-art bias mitigation methods by using three different fairness measurements. The results show that our approach successfully increases both accuracy and fairness for 61% of the cases studied, while the state-of-the-art always decrease accuracy when attempting to reduce bias. With our proposed approach, software engineers that previously were concerned with accuracy losses when considering fairness, are now enabled to improve the fairness of binary classification models without sacrificing accuracy.<\/jats:p>","DOI":"10.1007\/s10664-023-10419-3","type":"journal-article","created":{"date-parts":[[2024,1,3]],"date-time":"2024-01-03T10:02:16Z","timestamp":1704276136000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":12,"title":["Search-based Automatic Repair for Fairness and Accuracy in Decision-making Software"],"prefix":"10.1007","volume":"29","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8684-5909","authenticated-orcid":false,"given":"Max","family":"Hort","sequence":"first","affiliation":[]},{"given":"Jie M.","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Federica","family":"Sarro","sequence":"additional","affiliation":[]},{"given":"Mark","family":"Harman","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,3]]},"reference":[{"key":"10419_CR1","unstructured":"Agarwal A, Beygelzimer A, Dud\u00edk M, Langford J, Wallach H (2018) A reductions approach to fair classification. In: International conference on machine learning. PMLR, 60\u201369"},{"key":"10419_CR2","doi-asserted-by":"crossref","unstructured":"Aggarwal A, Lohia P, Nagar S, Dey K, Saha D (2019) Black box fairness testing of machine learning models. In: Proceedings of the 2019 27th ACM joint meeting on european software engineering conference and symposium on the foundations of software engineering. pp. 625\u2013635","DOI":"10.1145\/3338906.3338937"},{"key":"10419_CR3","doi-asserted-by":"crossref","unstructured":"Angell R, Johnson B, Brun Y, Meliou A. (2018) Themis: Automatically testing software for discrimination. In: Proceedings of the 2018 26th ACM joint meeting on european software engineering conference and symposium on the foundations of software engineering. pp. 871\u2013875","DOI":"10.1145\/3236024.3264590"},{"key":"10419_CR4","unstructured":"Angwin J, Larson J, Mattu S, Kirchner L (2016) Machine bias. ProPublica. See https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing\/"},{"issue":"3","key":"10419_CR5","first-page":"219","volume":"24","author":"A Arcuri","year":"2014","unstructured":"Arcuri A, Briand L (2014) A Hitchhiker\u2019s guide to statistical tests for assessing randomized algorithms in software engineering. STVR 24(3):219\u2013250","journal-title":"STVR"},{"key":"10419_CR6","unstructured":"Bellamy RKE, Dey K, Hind M, Hoffman SC, Houde S, Kannan K, Lohia P, Martino J, Mehta S, Mojsilovic A, et\u00a0al (2018) AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias. arXiv:1810.01943"},{"key":"10419_CR7","unstructured":"Berk R, Heidari H, Jabbari S, Joseph M, Kearns M, Morgenstern J, Neel S, Roth A (2017) A Convex Framework for Fair Regression. FAT-ML Workshop"},{"key":"10419_CR8","doi-asserted-by":"publisher","unstructured":"Berk R, Heidari H, Jabbari S, Kearns M, Roth A (2018) Fairness in criminal justice risk assessments: The state of the art. Sociological Methods & Research. https:\/\/doi.org\/10.1177\/0049124118782533","DOI":"10.1177\/0049124118782533"},{"key":"10419_CR9","doi-asserted-by":"crossref","unstructured":"Biswas S, Rajan H (2020) Do the Machine Learning Models on a Crowd Sourced Platform Exhibit Bias? An Empirical Study on Model Fairness. arXiv:2005.12379","DOI":"10.1145\/3368089.3409704"},{"key":"10419_CR10","volume-title":"Classification and regression trees","author":"L Breiman","year":"1984","unstructured":"Breiman L, Friedman J, Stone CJ, Olshen RA (1984) Classification and regression trees. CRC Press"},{"issue":"1","key":"10419_CR11","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1017\/S0269888997000015","volume":"12","author":"LA Breslow","year":"1997","unstructured":"Breslow LA, Aha DW (1997) Simplifying decision trees: A survey. Knowl Eng Rev 12(1):1\u201340","journal-title":"Knowl Eng Rev"},{"key":"10419_CR12","doi-asserted-by":"crossref","unstructured":"Brun Y, Meliou A (2018) Software fairness. In: Proceedings of the 2018 26th ACM joint meeting on european software engineering conference and symposium on the foundations of software engineering. 754\u2013759","DOI":"10.1145\/3236024.3264838"},{"key":"10419_CR13","doi-asserted-by":"crossref","unstructured":"Calders T, Kamiran F, Pechenizkiy M (2009) Building classifiers with independency constraints. In: 2009 IEEE international conference on data mining workshops. IEEE, 13\u201318","DOI":"10.1109\/ICDMW.2009.83"},{"key":"10419_CR14","doi-asserted-by":"crossref","unstructured":"Calders T, Karim A, Kamiran F, Ali W, Zhang X (2013) Controlling attribute effect in linear regression. In 2013 IEEE 13th international conference on data mining. IEEE, 71\u201380","DOI":"10.1109\/ICDM.2013.114"},{"issue":"2","key":"10419_CR15","doi-asserted-by":"publisher","first-page":"277","DOI":"10.1007\/s10618-010-0190-x","volume":"21","author":"T Calders","year":"2010","unstructured":"Calders T, Verwer S (2010) Three naive Bayes approaches for discrimination-free classification. Data Min Knowl Discov 21(2):277\u2013292","journal-title":"Data Min Knowl Discov"},{"key":"10419_CR16","unstructured":"Calmon F, Wei D, Vinzamuri B, Ramamurthy KN, Varshney KR (2017) Optimized pre-processing for discrimination prevention. In Advances in neural information processing systems. 3992\u20134001"},{"key":"10419_CR17","doi-asserted-by":"crossref","unstructured":"Celis LE, Huang L, Keswani V, Vishnoi NK (2019) Classification with fairness constraints: A meta-algorithm with provable guarantees. In: Proceedings of the conference on fairness, accountability, and transparency. 319\u2013328","DOI":"10.1145\/3287560.3287586"},{"key":"10419_CR18","doi-asserted-by":"crossref","unstructured":"Chakraborty J, Majumder S, Menzies T (2021) Bias in machine learning software: Why? how? what to do?. In: Proceedings of the 29th ACM joint meeting on european software engineering conference and symposium on the foundations of software engineering. 429\u2013440","DOI":"10.1145\/3468264.3468537"},{"key":"10419_CR19","doi-asserted-by":"crossref","unstructured":"Chakraborty J, Majumder S, Yu Z, Menzies T (2020) Fairway: a way to build fair ML software. In: Proceedings of the 28th ACM joint meeting on european software engineering conference and symposium on the foundations of software engineering. 654\u2013665","DOI":"10.1145\/3368089.3409697"},{"key":"10419_CR20","doi-asserted-by":"crossref","unstructured":"Chen J, Kallus N, Mao X, Svacha G, Udell M (2019) Fairness under unawareness: Assessing disparity when protected class is unobserved. In: Proceedings of the conference on fairness, accountability, and transparency. 339\u2013348","DOI":"10.1145\/3287560.3287594"},{"key":"10419_CR21","unstructured":"Chen Z, Zhang JM, Hort M, Sarro F, Harman M (2022a) Fairness Testing: A Comprehensive Survey and Analysis of Trends. arXiv:2207.10223"},{"key":"10419_CR22","doi-asserted-by":"crossref","unstructured":"Chen Z, Zhang JM, Sarro F, Harman M (2022b) MAAT: a novel ensemble approach to addressing fairness and performance bugs for machine learning software. In: Proceedings of the 30th ACM joint european software engineering conference and symposium on the foundations of software engineering. 1122\u20131134","DOI":"10.1145\/3540250.3549093"},{"issue":"4","key":"10419_CR23","doi-asserted-by":"publisher","first-page":"106:1","DOI":"10.1145\/3583561","volume":"32","author":"Z Chen","year":"2023","unstructured":"Chen Z, Zhang JM, Sarro F, Harman M (2023) A Comprehensive Empirical Study of Bias Mitigation Methods for Machine Learning Classifiers. ACM Trans Softw Eng Methodol 32(4):106:1-106:30","journal-title":"ACM Trans Softw Eng Methodol"},{"key":"10419_CR24","doi-asserted-by":"crossref","unstructured":"Chen Z, Zhang JM, Sarro F, Harman M (2023b) A Comprehensive Empirical Study of Bias Mitigation Methods for Machine Learning Classifiers. ACM Trans Softw Eng Methodol","DOI":"10.1145\/3583561"},{"key":"10419_CR25","doi-asserted-by":"crossref","unstructured":"Chen Z, Zhang JM, Sarro F, Harman M (2024) Fairness Improvement with Multiple Protected Attributes: How Far Are We?. In: International conference on software engineering (ICSE)","DOI":"10.1145\/3597503.3639083"},{"key":"10419_CR26","doi-asserted-by":"crossref","unstructured":"Corbett-Davies S, Pierson E, Feller A, Goel S, Huq A (2017)Algorithmic decision making and the cost of fairness. In: Proceedings of the 23rd ACM SIGKDD international conference on knowledge discovery and data mining. 797\u2013806","DOI":"10.1145\/3097983.3098095"},{"key":"10419_CR27","doi-asserted-by":"crossref","unstructured":"Dwork C, Hardt M, Pitassi T, Reingold O, Zemel R (2012) Fairness through awareness. In: Proceedings of the 3rd innovations in theoretical computer science conference. 214\u2013226","DOI":"10.1145\/2090236.2090255"},{"key":"10419_CR28","doi-asserted-by":"crossref","unstructured":"Feldman M, Friedler SA, Moeller J, Scheidegger C, Venkatasubramanian S (2015) Certifying and removing disparate impact. In: proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining. 259\u2013268","DOI":"10.1145\/2783258.2783311"},{"key":"10419_CR29","doi-asserted-by":"crossref","unstructured":"Ferrucci F, Gravino C, Oliveto R, Sarro F (2010) Genetic Programming for Effort Estimation: An Analysis of the Impact of Different Fitness Functions. In: 2nd International symposium on search based software engineering. 89\u201398","DOI":"10.1109\/SSBSE.2010.20"},{"key":"10419_CR30","doi-asserted-by":"crossref","unstructured":"Friedler SA, Scheidegger C, Venkatasubramanian S, Choudhary S, Hamilton EP, Roth D (2019) A comparative study of fairness-enhancing interventions in machine learning. In: Proceedings of the conference on fairness, accountability, and transparency. ACM, 329\u2013338","DOI":"10.1145\/3287560.3287589"},{"key":"10419_CR31","doi-asserted-by":"crossref","unstructured":"Galhotra S, Brun Y, Meliou A (2017) Fairness testing: testing software for discrimination. In: Proceedings of the 2017 11th joint meeting on foundations of software engineering. ACM, 498\u2013510","DOI":"10.1145\/3106237.3106277"},{"key":"10419_CR32","doi-asserted-by":"publisher","unstructured":"Gohar U, Cheng L (2023) A Survey on Intersectional Fairness in Machine Learning: Notions, Mitigation, and Challenges. In: Elkind E(Ed) Proceedings of the thirty-second international joint conference on artificial intelligence, IJCAI-23, . International Joint Conferences on Artificial Intelligence Organization, 6619\u20136627 https:\/\/doi.org\/10.24963\/ijcai.2023\/742","DOI":"10.24963\/ijcai.2023\/742"},{"key":"10419_CR33","unstructured":"Hardt M, Price E, Srebro N (2016) Equality of opportunity in supervised learning. In: Advances in neural information processing systems. 3315\u20133323"},{"key":"10419_CR34","doi-asserted-by":"crossref","unstructured":"Harman M, McMinn P, De\u00a0Souza JT, Yoo S (2010) Search based software engineering: Techniques, taxonomy, tutorial. In: Empirical software engineering and verification. Springer, 1\u201359","DOI":"10.1007\/978-3-642-25231-0_1"},{"key":"10419_CR35","doi-asserted-by":"crossref","unstructured":"Harrison G, Hanson J, Jacinto C, Ramirez J, Ur B (2020) An empirical study on the perceived fairness of realistic, imperfect machine learning models. In: Proceedings of the 2020 conference on fairness, accountability, and transparency. 392\u2013402","DOI":"10.1145\/3351095.3372831"},{"key":"10419_CR36","doi-asserted-by":"crossref","unstructured":"Horkoff J (2019) Non-functional requirements for machine learning: Challenges and new directions. In: 2019 IEEE 27th international requirements engineering conference (RE). IEEE, 386\u2013391","DOI":"10.1109\/RE.2019.00050"},{"key":"10419_CR37","doi-asserted-by":"crossref","unstructured":"Hort M, Chen Z, Zhang JM, Harman M, Sarro F (2023a) Bias mitigation for machine learning classifiers: A comprehensive survey. ACM J Resp Comput","DOI":"10.1145\/3631326"},{"key":"10419_CR38","doi-asserted-by":"crossref","unstructured":"Hort M, Chen Z, Zhang JM, Harman M, Sarro F (2023b) Bias Mitigation for Machine Learning Classifiers: A Comprehensive Survey. ACM J Resp Comput. arXiv:2207.07068","DOI":"10.1145\/3631326"},{"issue":"109916","key":"10419_CR39","doi-asserted-by":"publisher","first-page":"1568","DOI":"10.1016\/j.asoc.2022.109916","volume":"133","author":"M Hort","year":"2023","unstructured":"Hort M, Moussa R, Sarro F (2023) Multi-objective search for gender-fair and semantically correct word embeddings. Appl Soft Comput 133(109916):1568\u20134946. https:\/\/doi.org\/10.1016\/j.asoc.2022.109916","journal-title":"Appl Soft Comput"},{"key":"10419_CR40","doi-asserted-by":"crossref","unstructured":"Hort M, Sarro F (2021) Did you do your homework? Raising awareness on software fairness and discrimination. In: 2021 36th IEEE\/ACM international conference on automated software engineering (ASE). IEEE, 1322\u20131326","DOI":"10.1109\/ASE51524.2021.9678568"},{"key":"10419_CR41","doi-asserted-by":"crossref","unstructured":"Hort M, Zhang JM, Harman M, Sarro F (2023d) On-line Appendix to the article Search-based Automatic Repair for Fairness and Accuracy in Decision-making Software https:\/\/github.com\/SOLAR-group\/Fairness-Postprocessing","DOI":"10.1007\/s10664-023-10419-3"},{"key":"10419_CR42","doi-asserted-by":"crossref","unstructured":"Hort M, Zhang JM, Sarro F, Harman M (2021) Fairea: A model behaviour mutation approach to benchmarking bias mitigation methods. In: Proceedings of the 29th ACM joint meeting on european software engineering conference and symposium on the foundations of software engineering. 994\u20131006","DOI":"10.1145\/3468264.3468565"},{"key":"10419_CR43","doi-asserted-by":"publisher","unstructured":"Jacobs AZ, Wallach H (2021) Measurement and Fairness. In: Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (FAccT \u201921). Association for Computing Machinery, New York, NY, USA, 375-385. 9781450383097 https:\/\/doi.org\/10.1145\/3442188.3445901","DOI":"10.1145\/3442188.3445901"},{"key":"10419_CR44","doi-asserted-by":"crossref","unstructured":"Kamiran F, Calders T (2009) Classifying without discriminating. In: 2009 2nd international conference on computer, control and communication. IEEE, 1\u20136","DOI":"10.1109\/IC4.2009.4909197"},{"issue":"1","key":"10419_CR45","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10115-011-0463-8","volume":"33","author":"F Kamiran","year":"2012","unstructured":"Kamiran F, Calders T (2012) Data preprocessing techniques for classification without discrimination. Knowl Inf Syst 33(1):1\u201333","journal-title":"Knowl Inf Syst"},{"key":"10419_CR46","doi-asserted-by":"crossref","unstructured":"Kamiran F, Calders T, Pechenizkiy M (2010) Discrimination aware decision tree learning. In: 2010 IEEE international conference on data mining. IEEE, 869\u2013874","DOI":"10.1109\/ICDM.2010.50"},{"key":"10419_CR47","doi-asserted-by":"crossref","unstructured":"Kamiran F, Karim A, Zhang X (2012) Decision theory for discrimination-aware classification. In: 2012 IEEE 12th international conference on data mining. IEEE, 924\u2013929","DOI":"10.1109\/ICDM.2012.45"},{"key":"10419_CR48","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1016\/j.ins.2017.09.064","volume":"425","author":"F Kamiran","year":"2018","unstructured":"Kamiran F, Mansha S, Karim A, Zhang X (2018) Exploiting reject option in classification for social discrimination control. Inf Sci 425:18\u201333","journal-title":"Inf Sci"},{"key":"10419_CR49","doi-asserted-by":"crossref","unstructured":"Kamishima T, Akaho S, Asoh H, Sakuma J (2012) Fairness-aware classifier with prejudice remover regularizer. In: Joint european conference on machine learning and knowledge discovery in databases. Springer, 35\u201350","DOI":"10.1007\/978-3-642-33486-3_3"},{"key":"10419_CR50","unstructured":"Kearns M, Neel S, Roth A, Wu ZS (2018) Preventing Fairness Gerrymandering: Auditing and Learning for Subgroup Fairness. In: Dy J, Krause A (Eds) Proceedings of Machine Learning Research, Vol.\u00a080. PMLR, Stockholmsm\u00e4ssan, Stockholm Sweden, 2564\u20132572. http:\/\/proceedings.mlr.press\/v80\/kearns18a.html"},{"issue":"6","key":"10419_CR51","first-page":"1403","volume":"38","author":"E Kocaguneli","year":"2011","unstructured":"Kocaguneli E, Menzies T, Keung JW (2011) On the value of ensemble effort estimation. IEEE TSE 38(6):1403\u20131416","journal-title":"IEEE TSE"},{"key":"10419_CR52","unstructured":"Kohav R (2023) Adult data set. http:\/\/archive.ics.uci.edu\/ml\/datasets\/adult"},{"key":"10419_CR53","unstructured":"Li X, Chen Z, Zhang JM, Sarro F, Zhang Y, Liu X (2023) Dark-Skin Individuals Are at More Risk on the Street: Unmasking Fairness Issues of Autonomous Driving Systems. arXiv:abs\/2308.02935"},{"key":"10419_CR54","unstructured":"Ma M, Tian Z, Hort M, Sarro F, Zhang H, Lin Q, Zhang D (2022) Enhanced Fairness Testing via Generating Effective Initial Individual Discriminatory Instances. arXiv:cs.SE\/2209.08321"},{"key":"10419_CR55","unstructured":"Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A (2019) A survey on bias and fairness in machine learning. arXiv:1908.09635"},{"key":"10419_CR56","doi-asserted-by":"crossref","unstructured":"Mikians J, Gyarmati L, Erramilli V, Laoutaris N (2012) Detecting price and search discrimination on the internet. In: Proceedings of the 11th ACM workshop on hot topics in networks. 79\u201384","DOI":"10.1145\/2390231.2390245"},{"key":"10419_CR57","doi-asserted-by":"publisher","first-page":"22","DOI":"10.1016\/j.dss.2014.03.001","volume":"62","author":"S Moro","year":"2014","unstructured":"Moro S, Cortez P, Rita P (2014) A data-driven approach to predict the success of bank telemarketing. Decis Support Syst 62:22\u201331","journal-title":"Decis Support Syst"},{"key":"10419_CR58","doi-asserted-by":"crossref","unstructured":"Moussa R, Sarro F (2022) On the Use of Evaluation Measures for Defect Prediction Studies. In: 2022 ACM SIGSOFT international symposium on software testing and analysis (ISSTA. ACM","DOI":"10.1145\/3533767.3534405"},{"key":"10419_CR59","first-page":"2825","volume":"12","author":"F Pedregosa","year":"2011","unstructured":"Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, Blondel M, Prettenhofer P, Weiss R, Dubourg V, Vanderplas J, Passos A, Cournapeau D, Brucher M, Perrot M, Duchesnay E (2011) Scikit-learn: Machine Learning in Python. J Mach Learn Res 12:2825\u20132830","journal-title":"J Mach Learn Res"},{"key":"10419_CR60","doi-asserted-by":"crossref","unstructured":"Pedreshi D, Ruggieri S, Turini F (2008) Discrimination-aware data mining. In: Proceedings of the 14th ACM SIGKDD international conference on Knowledge discovery and data mining. 560\u2013568","DOI":"10.1145\/1401890.1401959"},{"issue":"3","key":"10419_CR61","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1007\/s10664-022-10116-7","volume":"27","author":"A Perera","year":"2022","unstructured":"Perera A, Aleti A, Tantithamthavorn C, Jiarpakdee J, Turhan B, Kuhn L, Walker K (2022) Search-Based Fairness Testing for Regression-Based Machine Learning Systems. Empir Softw Eng 27(3):79. https:\/\/doi.org\/10.1007\/s10664-022-10116-7","journal-title":"Empir Softw Eng"},{"issue":"3","key":"10419_CR62","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3494672","volume":"55","author":"D Pessach","year":"2022","unstructured":"Pessach D, Shmueli E (2022) A review on fairness in machine learning. ACM Comput Surv (CSUR) 55(3):1\u201344","journal-title":"ACM Comput Surv (CSUR)"},{"key":"10419_CR63","unstructured":"Pleiss G, Raghavan M, Wu F, Kleinberg J, Weinberger KQ (2017) On fairness and calibration. In: Advances in neural information processing systems. 5680\u20135689"},{"key":"10419_CR64","unstructured":"propublica (2023) data for the propublica story \u2018machine bias\u2019. https:\/\/github.com\/propublica\/compas-analysis\/"},{"issue":"3","key":"10419_CR65","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1016\/S0020-7373(87)80053-6","volume":"27","author":"JR Quinlan","year":"1987","unstructured":"Quinlan JR (1987) Simplifying decision trees. Int J Man-Mach Stud 27(3):221\u2013234","journal-title":"Int J Man-Mach Stud"},{"key":"10419_CR66","unstructured":"Romei A, Ruggieri S (2011) A multidisciplinary survey on discrimination analysis"},{"key":"10419_CR67","doi-asserted-by":"crossref","unstructured":"Sarro F (2023) Search-based software engineering in the era of modern software systems. In: Procs. of the 31st IEEE international requirements engineering conferece","DOI":"10.1109\/RE57278.2023.00010"},{"issue":"10","key":"10419_CR68","first-page":"898","volume":"43","author":"F Sarro","year":"2017","unstructured":"Sarro F, Ferrucci F, Harman M, Manna A, Ren J (2017) Adaptive Multi-Objective Evolutionary Algorithms for Overtime Planning in Software Projects. IEEE TSE 43(10):898\u2013917","journal-title":"IEEE TSE"},{"key":"10419_CR69","doi-asserted-by":"crossref","unstructured":"Sarro F, Harman M, Jia Y, Zhang Y (2018) Customer rating reactions can be predicted purely using app features. In: IEEE international requirements engineering conference. 76\u201387","DOI":"10.1109\/RE.2018.00018"},{"issue":"3","key":"10419_CR70","doi-asserted-by":"publisher","first-page":"12:1","DOI":"10.1145\/3234940","volume":"27","author":"F Sarro","year":"2018","unstructured":"Sarro F, Petrozziello A (2018) Linear Programming As a Baseline for Software Effort Estimation. ACM TOSEM 27(3):12:1-12:28","journal-title":"ACM TOSEM"},{"key":"10419_CR71","doi-asserted-by":"crossref","unstructured":"Sarro F, Petrozziello A, Harman M. (2016). Multi-objective software effort estimation. In Procs. of the international conference on software engineering (ICSE). IEEE, 619\u2013630","DOI":"10.1145\/2884781.2884830"},{"issue":"2020","key":"10419_CR72","first-page":"2798","volume":"33","author":"Y Savani","year":"2020","unstructured":"Savani Y, White C, Govindarajulu NS (2020) Intra-processing methods for debiasing neural networks. Adv Neural Inf Process 33(2020):2798\u20132810","journal-title":"Adv Neural Inf Process"},{"key":"10419_CR73","doi-asserted-by":"crossref","unstructured":"Speicher T, Heidari H, Grgic-Hlaca N, Gummadi KP, Singla A, Weller A, Zafar MB (2018) A unified approach to quantifying algorithmic unfairness: Measuring individual & group unfairness via inequality indices. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining. 2239\u20132248","DOI":"10.1145\/3219819.3220046"},{"key":"10419_CR74","doi-asserted-by":"crossref","unstructured":"Tizpaz-Niari S, Kumar A, Tan G, Trivedi A (2022) Fairness-aware configuration of machine learning libraries. arXiv:2202.06196","DOI":"10.1145\/3510003.3510202"},{"key":"10419_CR75","doi-asserted-by":"crossref","unstructured":"Udeshi S, Arora P, Chattopadhyay S (2018) Automated directed fairness testing. In Proceedings of the 33rd ACM\/IEEE international conference on automated software engineering. 98\u2013108","DOI":"10.1145\/3238147.3238165"},{"issue":"2","key":"10419_CR76","first-page":"101","volume":"25","author":"A Vargha","year":"2000","unstructured":"Vargha A, Delaney HD (2000) A critique and improvement of the CL common language effect size statistics of McGraw and Wong. J Educ Behav Stat 25(2):101\u2013132","journal-title":"J Educ Behav Stat"},{"key":"10419_CR77","doi-asserted-by":"crossref","unstructured":"Wilcoxon F (1992) Individual comparisons by ranking methods. In: Breakthroughs in statistics. Springer, 196\u2013202","DOI":"10.1007\/978-1-4612-4380-9_16"},{"key":"10419_CR78","unstructured":"Zafar MB, Valera I, Rogriguez MG, Gummadi KP (2017) Fairness constraints: Mechanisms for fair classification. In: Artificial intelligence and statistics. 962\u2013970"},{"key":"10419_CR79","unstructured":"Zemel R, Wu Y, Swersky K, Pitassi T, Dwork C (2013) Learning fair representations. In: International conference on machine learning. 325\u2013333"},{"key":"10419_CR80","doi-asserted-by":"crossref","unstructured":"Zhang BH, Lemoine B, Mitchell M (2018) Mitigating unwanted biases with adversarial learning. In: Proceedings of the 2018 AAAI\/ACM conference on ai, ethics, and society. ACM, 335\u2013340","DOI":"10.1145\/3278721.3278779"},{"key":"10419_CR81","doi-asserted-by":"crossref","unstructured":"Zhang J, Harman M (2021) Ignorance and Prejudice in Software Fairness. In: 2021 IEEE\/ACM 43th international conference on software engineering (ICSE). IEEE","DOI":"10.1109\/ICSE43902.2021.00129"},{"key":"10419_CR82","doi-asserted-by":"crossref","unstructured":"Zhang JM, Harman M, Ma L, Liu Y (2020) Machine Learning Testing: Survey, Landscapes and Horizons. IEEE Trans Softw Eng 1(1)","DOI":"10.1109\/TSE.2019.2962027"},{"key":"10419_CR83","doi-asserted-by":"crossref","unstructured":"\u017dliobaite I, Kamiran F, Calders T (2011) Handling conditional discrimination. In: 2011 IEEE 11th international conference on data mining. IEEE, 992\u20131001","DOI":"10.1109\/ICDM.2011.72"}],"container-title":["Empirical Software Engineering"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10664-023-10419-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10664-023-10419-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10664-023-10419-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,7]],"date-time":"2024-11-07T08:37:18Z","timestamp":1730968638000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10664-023-10419-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1]]},"references-count":83,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2024,1]]}},"alternative-id":["10419"],"URL":"https:\/\/doi.org\/10.1007\/s10664-023-10419-3","relation":{},"ISSN":["1382-3256","1573-7616"],"issn-type":[{"value":"1382-3256","type":"print"},{"value":"1573-7616","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,1]]},"assertion":[{"value":"25 October 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 January 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}],"article-number":"36"}}