{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T22:42:19Z","timestamp":1775169739853,"version":"3.50.1"},"reference-count":50,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2022,3,7]],"date-time":"2022-03-07T00:00:00Z","timestamp":1646611200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,3,7]],"date-time":"2022-03-07T00:00:00Z","timestamp":1646611200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["AI Ethics"],"published-print":{"date-parts":[[2023,2]]},"DOI":"10.1007\/s43681-022-00147-7","type":"journal-article","created":{"date-parts":[[2022,3,7]],"date-time":"2022-03-07T14:02:47Z","timestamp":1646661767000},"page":"267-279","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":38,"title":["Fairness Score and process standardization: framework for fairness certification in artificial intelligence systems"],"prefix":"10.1007","volume":"3","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4553-5861","authenticated-orcid":false,"given":"Avinash","family":"Agarwal","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3193-1521","authenticated-orcid":false,"given":"Harsh","family":"Agarwal","sequence":"additional","affiliation":[]},{"given":"Nihaarika","family":"Agarwal","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,3,7]]},"reference":[{"key":"147_CR1","doi-asserted-by":"crossref","unstructured":"Sharma, S., Henderson, J., Ghosh, J.: Certifai: counterfactual explanations for robustness, transparency, interpretability, and fairness of artificial intelligence models (2019). arXiv:1905.07857","DOI":"10.1145\/3375627.3375812"},{"key":"147_CR2","doi-asserted-by":"crossref","unstructured":"Yapo, A., Weiss, J.W.: Ethical implications of bias in machine learning. In: HICSS (2018)","DOI":"10.24251\/HICSS.2018.668"},{"key":"147_CR3","unstructured":"Curionin, A.: Trusted AI for Business. https:\/\/docbox.etsi.org\/Workshop\/2019\/201904_SUMMIT_AI\/S04_AI_CHALLENGES_OPPORT\/TRUSTED_AI_BUSINESS_IBM_CURIONI.pdfAccessed 14 Aug 2020"},{"key":"147_CR4","unstructured":"Mahoney, T., Varshney, K.R., Hind, M.: AI Fairness. O\u2019Reilly Media, Inc. (2020)"},{"key":"147_CR5","unstructured":"Kodiyan, A.A.: An overview of ethical issues in using AI systems in hiring with a case study of Amazon\u2019s AI based hiring tool. Researchgate Preprint (2019)"},{"key":"147_CR6","doi-asserted-by":"crossref","unstructured":"Hann\u00e1k, A., Wagner, C., Garcia, D., Mislove, A., Strohmaier, M., Wilson, C.: Bias in online freelance marketplaces: evidence from taskrabbit and fiverr. In: Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing, pp. 1914\u20131933 (2017)","DOI":"10.1145\/2998181.2998327"},{"key":"147_CR7","unstructured":"Garvie, C., Frankle, J.: Facial-recognition software might have a racial bias problem. Atlantic 7 (2016). https:\/\/www.theatlantic.com\/technology\/archive\/2016\/04\/the-underlying-bias-of-facial-recognition-systems\/476991\/"},{"issue":"6464","key":"147_CR8","doi-asserted-by":"publisher","first-page":"447","DOI":"10.1126\/science.aax2342","volume":"366","author":"Z Obermeyer","year":"2019","unstructured":"Obermeyer, Z., Powers, B., Vogeli, C., Mullainathan, S.: Dissecting racial bias in an algorithm used to manage the health of populations. Science 366(6464), 447\u2013453 (2019)","journal-title":"Science"},{"key":"147_CR9","unstructured":"Wadsworth, C., Vera, F., Piech, C.: Achieving fairness through adversarial learning: an application to recidivism prediction (2018). arXiv:1807.00199"},{"key":"147_CR10","unstructured":"Hadhazy, A.: Biased bots: artificial-intelligence systems echo human prejudices. Princeton University (2017)"},{"issue":"5","key":"147_CR11","doi-asserted-by":"publisher","first-page":"44","DOI":"10.1145\/2447976.2447990","volume":"56","author":"L Sweeney","year":"2013","unstructured":"Sweeney, L.: Discrimination in online ad delivery. Commun. ACM 56(5), 44\u201354 (2013)","journal-title":"Commun. ACM"},{"key":"147_CR12","unstructured":"Lee, N.T., Resnick, P., Barton, G.: Algorithmic bias detection and mitigation: best practices and policies to reduce consumer harms. Brookings Institute: Washington, DC (2019)"},{"key":"147_CR13","unstructured":"Leavy, S., O\u2019Sullivan, B., Siapera, E.: Data, power and bias in artificial intelligence (2020). arXiv:2008.07341"},{"key":"147_CR14","doi-asserted-by":"crossref","unstructured":"Zheng, K., Gao, J., Ngiam, K.Y., Ooi, B.C., Yip, W.L.J.: Resolving the bias in electronic medical records. In: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 2171\u20132180 (2017)","DOI":"10.1145\/3097983.3098149"},{"key":"147_CR15","doi-asserted-by":"crossref","unstructured":"Segal, S., Adi, Y., Pinkas, B., Baum, C., Ganesh, C., Keshet, J.: Fairness in the eyes of the data: certifying machine-learning models. In: Proceedings of the 2021 AAAI\/ACM Conference on AI, Ethics, and Society, pp. 926\u2013935 (2021)","DOI":"10.1145\/3461702.3462554"},{"key":"147_CR16","unstructured":"Tommasi, T., Bucci, S., Caputo, B., Asinari, P.: Towards fairness certification in artificial intelligence (2021). arXiv:2106.02498"},{"issue":"1","key":"147_CR17","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1177\/2031952519829082","volume":"10","author":"C Schubert","year":"2019","unstructured":"Schubert, C., H\u00fctt, M.T.: Economy-on-demand and the fairness of algorithms. Eur. Labour Law J. 10(1), 3\u201316 (2019)","journal-title":"Eur. Labour Law J."},{"key":"147_CR18","unstructured":"Chouldechova, A., Roth, A.: The frontiers of fairness in machine learning (2018). arXiv:1810.08810"},{"issue":"2","key":"147_CR19","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1109\/MTS.2021.3056293","volume":"40","author":"X Ferrer","year":"2021","unstructured":"Ferrer, X., van Nuenen, T., Such, J.M., Cot\u00e9, M., Criado, N.: Bias and discrimination in AI: a cross-disciplinary perspective. IEEE Technol. Soc. Mag. 40(2), 72\u201380 (2021)","journal-title":"IEEE Technol. Soc. Mag."},{"issue":"2128","key":"147_CR20","doi-asserted-by":"publisher","first-page":"20170362","DOI":"10.1098\/rsta.2017.0362","volume":"376","author":"H Shah","year":"2018","unstructured":"Shah, H.: Algorithmic accountability. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 376(2128), 20170362 (2018)","journal-title":"Philos. Trans. R. Soc. A Math. Phys. Eng. Sci."},{"key":"147_CR21","doi-asserted-by":"publisher","unstructured":"Panch, T., Mattie, H., Atun, R.: Artificial intelligence and algorithmic bias: implications for health systems. J. Glob. Health 9(2) (2019). https:\/\/doi.org\/10.7189\/jogh.09.020318","DOI":"10.7189\/jogh.09.020318"},{"issue":"3","key":"147_CR22","doi-asserted-by":"publisher","first-page":"e1356","DOI":"10.1002\/widm.1356","volume":"10","author":"E Ntoutsi","year":"2020","unstructured":"Ntoutsi, E., Fafalios, P., Gadiraju, U., Iosifidis, V., Nejdl, W., Vidal, M.E., et al.: Bias in data-driven artificial intelligence systems\u2014an introductory survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 10(3), e1356 (2020)","journal-title":"Wiley Interdiscip. Rev. Data Min. Knowl. Discov."},{"issue":"3","key":"147_CR23","doi-asserted-by":"publisher","first-page":"330","DOI":"10.1145\/230538.230561","volume":"14","author":"B Friedman","year":"1996","unstructured":"Friedman, B., Nissenbaum, H.: Bias in computer systems. ACM Trans. Inf. Syst. (TOIS) 14(3), 330\u2013347 (1996)","journal-title":"ACM Trans. Inf. Syst. (TOIS)"},{"key":"147_CR24","doi-asserted-by":"crossref","unstructured":"Hajian, S., Bonchi, F., Castillo, C.: Algorithmic bias: from discrimination discovery to fairness-aware data mining. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pp. 2125\u20132126 (2016)","DOI":"10.1145\/2939672.2945386"},{"key":"147_CR25","doi-asserted-by":"crossref","unstructured":"Verma, S., Rubin, J.: Fairness definitions explained. In: IEEE\/ACM International Workshop on Software Fairness (fairware), pp 1\u20137. IEEE (2018)","DOI":"10.1145\/3194770.3194776"},{"key":"147_CR26","unstructured":"Zliobaite, I.: Fairness-aware machine learning: a perspective (2017). arXiv:1708.00754"},{"issue":"6","key":"147_CR27","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3457607","volume":"54","author":"N Mehrabi","year":"2021","unstructured":"Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., Galstyan, A.: A survey on bias and fairness in machine learning. ACM Comput. Surv. (CSUR) 54(6), 1\u201335 (2021)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"147_CR28","doi-asserted-by":"crossref","unstructured":"Hutchinson, B., Mitchell, M.: 50 years of test (un) fairness: lessons for machine learning. In: Proceedings of the Conference on Fairness, Accountability, and Transparency, pp. 49\u201358 (2019)","DOI":"10.1145\/3287560.3287600"},{"issue":"4\/5","key":"147_CR29","doi-asserted-by":"publisher","first-page":"6-1","DOI":"10.1147\/JRD.2019.2942288","volume":"63","author":"M Arnold","year":"2019","unstructured":"Arnold, M., Bellamy, R.K., Hind, M., Houde, S., Mehta, S., Mojsilovi\u0107, A., et al.: FactSheets: increasing trust in AI services through supplier\u2019s declarations of conformity. IBM J. Res. Dev. 63(4\/5), 6\u20131 (2019)","journal-title":"IBM J. Res. Dev."},{"issue":"CSCW","key":"147_CR30","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3359221","volume":"3","author":"DK Mulligan","year":"2019","unstructured":"Mulligan, D.K., Kroll, J.A., Kohli, N., Wong, R.Y.: This thing called fairness: disciplinary confusion realizing a value in technology. Proc. ACM Hum. Comput. Interact. 3(CSCW), 1\u201336 (2019)","journal-title":"Proc. ACM Hum. Comput. Interact."},{"key":"147_CR31","unstructured":"Diakopoulos, N., Friedler, S., Arenas, M., Barocas, S., Hay, M., Howe, B., et\u00a0al.: Principles for accountable algorithms and a social impact statement for algorithms (2017). https:\/\/www.fatml.org\/resources\/principles-for-accountable-algorithms"},{"key":"147_CR32","doi-asserted-by":"crossref","unstructured":"Aggarwal, A., Lohia, P., Nagar, S., Dey, K., Saha, D.: Black box fairness testing of machine learning models. In: Proceedings of the 2019 27th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, pp. 625\u2013635 (2019)","DOI":"10.1145\/3338906.3338937"},{"key":"147_CR33","doi-asserted-by":"crossref","unstructured":"Dwork, C., Hardt, M., Pitassi, T., Reingold, O., Zemel, R.: Fairness through awareness. In: Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, pp. 214\u2013226 (2012)","DOI":"10.1145\/2090236.2090255"},{"key":"147_CR34","unstructured":"Kusner, M.J., Loftus, J.R., Russell, C., Silva, R.: Counterfactual fairness (2017). arXiv:1703.06856"},{"key":"147_CR35","unstructured":"Corbett-Davies, S., Goel, S.: The measure and mismeasure of fairness: a critical review of fair machine learning (2018). arXiv:1808.00023"},{"issue":"5","key":"147_CR36","doi-asserted-by":"publisher","first-page":"582","DOI":"10.1017\/S0269888913000039","volume":"29","author":"A Romei","year":"2014","unstructured":"Romei, A., Ruggieri, S.: A multidisciplinary survey on discrimination analysis. Knowl. Eng. Rev. 29(5), 582\u2013638 (2014)","journal-title":"Knowl. Eng. Rev."},{"key":"147_CR37","doi-asserted-by":"crossref","unstructured":"Hertweck, C., Heitz, C., Loi, M.: On the moral justification of statistical parity. In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pp. 747\u2013757 (2021)","DOI":"10.1145\/3442188.3445936"},{"key":"147_CR38","first-page":"671","volume":"104","author":"S Barocas","year":"2016","unstructured":"Barocas, S., Selbst, A.D.: Big data\u2019s disparate impact. Calif. Law Rev. 104, 671 (2016)","journal-title":"Calif. Law Rev."},{"key":"147_CR39","unstructured":"Bellamy, R.K., Dey, K., Hind, M., Hoffman, S.C., Houde, S., Kannan, K., et\u00a0al.: AI Fairness 360: an extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias (2018). arXiv:1810.01943"},{"key":"147_CR40","unstructured":"Kleinberg, J., Mullainathan, S., Raghavan, M.: Inherent trade-offs in the fair determination of risk scores (2016). arXiv:1609.05807"},{"key":"147_CR41","doi-asserted-by":"crossref","unstructured":"Singh, H., Singh, R., Mhasawade, V., Chunara, R.: Fairness violations and mitigation under covariate shift. In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pp. 3\u201313 (2021)","DOI":"10.1145\/3442188.3445865"},{"key":"147_CR42","unstructured":"Hinnefeld, J.H., Cooman, P., Mammo, N., Deese, R.: Evaluating fairness metrics in the presence of dataset bias (2018). arXiv:1809.09245"},{"key":"147_CR43","doi-asserted-by":"crossref","unstructured":"Lohia, P.K., Ramamurthy, K.N., Bhide, M., Saha, D., Varshney, K.R., Puri, R.: Bias mitigation post-processing for individual and group fairness. In: Icassp 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (icassp), pp. 2847\u20132851. IEEE (2019)","DOI":"10.1109\/ICASSP.2019.8682620"},{"key":"147_CR44","first-page":"3315","volume":"29","author":"M Hardt","year":"2016","unstructured":"Hardt, M., Price, E., Srebro, N.: Equality of opportunity in supervised learning. Adv. Neural. Inf. Process. Syst. 29, 3315\u20133323 (2016)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"147_CR45","unstructured":"Awwad, Y., Fletcher, R., Frey, D., Gandhi, A., Najafian, M., Teodorescu, M.: Exploring fairness in machine learning for international development. CITE MIT D-Lab (2020)"},{"key":"147_CR46","doi-asserted-by":"crossref","unstructured":"Corbett-Davies, S., Pierson, E., Feller, A., Goel, S., Huq, A.: Algorithmic decision making and the cost of fairness. In: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 797\u2013806 (2017)","DOI":"10.1145\/3097983.3098095"},{"key":"147_CR47","doi-asserted-by":"crossref","unstructured":"Castelnovo, A., Crupi, R., Greco, G., Regoli, D.: The zoo of fairness metrics in machine learning (2021). arXiv:2106.00467","DOI":"10.21203\/rs.3.rs-1162350\/v1"},{"issue":"1","key":"147_CR48","doi-asserted-by":"publisher","first-page":"29","DOI":"10.7815\/ijorcs.21.2011.011","volume":"2","author":"S Pandit","year":"2011","unstructured":"Pandit, S., Gupta, S., et al.: A comparative study on distance measuring approaches for clustering. Int. J. Res. Comput. Sci. 2(1), 29\u201331 (2011)","journal-title":"Int. J. Res. Comput. Sci."},{"key":"147_CR49","unstructured":"Tram\u00e8r, F., Atlidakis, V., Geambasu, R., Hsu, D.J., Hubaux, J.P., Humbert, M., et\u00a0al.: Discovering unwarranted associations in data-driven applications with the fairtest testing toolkit (2015). arXiv:1510.02377 [CoRR abs]"},{"key":"147_CR50","unstructured":"CFR.: Section 4D, Part 1607\u2014uniform guidelines on employee selection procedures. Code of Federal Regulations (1978)"}],"container-title":["AI and Ethics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-022-00147-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s43681-022-00147-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-022-00147-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,23]],"date-time":"2023-02-23T07:10:57Z","timestamp":1677136257000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s43681-022-00147-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,3,7]]},"references-count":50,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2023,2]]}},"alternative-id":["147"],"URL":"https:\/\/doi.org\/10.1007\/s43681-022-00147-7","relation":{},"ISSN":["2730-5953","2730-5961"],"issn-type":[{"value":"2730-5953","type":"print"},{"value":"2730-5961","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,3,7]]},"assertion":[{"value":"11 January 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 February 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 March 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"On behalf of all authors, the corresponding author states that there is no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Code availability"}}]}}