{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T00:40:29Z","timestamp":1776127229318,"version":"3.50.1"},"reference-count":37,"publisher":"IBM","issue":"4\/5","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IBM J. Res. &amp; Dev."],"published-print":{"date-parts":[[2019,7,1]]},"DOI":"10.1147\/jrd.2019.2942287","type":"journal-article","created":{"date-parts":[[2019,9,18]],"date-time":"2019-09-18T16:08:35Z","timestamp":1568822915000},"page":"4:1-4:15","source":"Crossref","is-referenced-by-count":647,"title":["AI Fairness 360: An extensible toolkit for detecting and mitigating algorithmic bias"],"prefix":"10.1147","volume":"63","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9403-2913","authenticated-orcid":false,"given":"R. K. E.","family":"Bellamy","sequence":"first","affiliation":[]},{"given":"K.","family":"Dey","sequence":"additional","affiliation":[]},{"given":"M.","family":"Hind","sequence":"additional","affiliation":[]},{"given":"S. C.","family":"Hoffman","sequence":"additional","affiliation":[]},{"given":"S.","family":"Houde","sequence":"additional","affiliation":[]},{"given":"K.","family":"Kannan","sequence":"additional","affiliation":[]},{"given":"P.","family":"Lohia","sequence":"additional","affiliation":[]},{"given":"J.","family":"Martino","sequence":"additional","affiliation":[]},{"given":"S.","family":"Mehta","sequence":"additional","affiliation":[]},{"given":"A.","family":"Mojsilovi\u0107","sequence":"additional","affiliation":[]},{"given":"S.","family":"Nagar","sequence":"additional","affiliation":[]},{"given":"K. Natesan","family":"Ramamurthy","sequence":"additional","affiliation":[]},{"given":"J.","family":"Richards","sequence":"additional","affiliation":[]},{"given":"D.","family":"Saha","sequence":"additional","affiliation":[]},{"given":"P.","family":"Sattigeri","sequence":"additional","affiliation":[]},{"given":"M.","family":"Singh","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7376-5536","authenticated-orcid":false,"given":"K. R.","family":"Varshney","sequence":"additional","affiliation":[]},{"given":"Y.","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"3082","reference":[{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1089\/big.2016.0048"},{"key":"ref32","article-title":"A large-scale analysis of racial disparities in police stops across the United States","author":"pierson","year":"2019","journal-title":"arXiv 1706 05678"},{"key":"ref31","first-page":"841","article-title":"Counterfactual explanations without opening the black box: Automated decisions and the GDPR","volume":"31","author":"wachter","year":"2018","journal-title":"Harvard J Law Technol"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287566"},{"key":"ref37","article-title":"Welfare and distributional impacts of fair classification","author":"hu","year":"0","journal-title":"Fairness Accountability and Transparency in Machine Learning"},{"key":"ref36","first-page":"5680","article-title":"On fairness and calibration","author":"pleiss","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref35","first-page":"3315","article-title":"Equality of opportunity in supervised learning","author":"hardt","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1145\/3278721.3278779"},{"key":"ref10","article-title":"Aequitas: Bias and fairness audit","author":"stevens","year":"2018"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1145\/3106237.3106277"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1080\/15228835.2017.1416512"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/s10115-011-0463-8"},{"key":"ref14","first-page":"4069","article-title":"Counterfactual fairness","author":"kusner","year":"2017","journal-title":"Proc 31st Int Conf Neural Inf Process Syst"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2012.45"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1145\/2783258.2783311"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-33486-3_3"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/s10618-010-0190-x"},{"key":"ref19","article-title":"Scaling up the accuracy of Naive&#x2013;Bayes classifiers: A decision-tree hybrid","author":"kohavi","year":"0","journal-title":"Proc Int'l Conf Knowledge Discovery and Data Mining"},{"key":"ref28","article-title":"On the (im)possibility of fairness","author":"friedler","year":"2016","journal-title":"arXiv 1609 07236"},{"key":"ref4","article-title":"Technical response to Northpointe","author":"larson","year":"2016"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220046"},{"key":"ref3","article-title":"COMPAS risk scales: Demonstrating accuracy equity and predictive parity","author":"dieterich","year":"2016"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287589"},{"key":"ref29","article-title":"AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias","author":"bellamy","year":"2018","journal-title":"arXiv 1810 01943"},{"key":"ref5","first-page":"43.1","article-title":"Inherent trade-offs in the fair determination of risk scores","author":"kleinberg","year":"2017","journal-title":"Proc Innov Theor Comput Sci Conf"},{"key":"ref8","article-title":"FairML: Toolbox for diagnosing bias in predictive modeling","author":"adebayo","year":"2016"},{"key":"ref7","article-title":"Fairness measures: Datasets and software for detecting algorithmic discrimination","author":"zehlike","year":"2017"},{"key":"ref2","article-title":"How we analyzed the COMPAS recidivism algorithm","author":"larson","year":"2016"},{"key":"ref9","first-page":"401","article-title":"FairTest: Discovering unwarranted associations in data-driven applications","author":"tram\u00e8r","year":"0","journal-title":"Proc IEEE Symp Privacy Secur"},{"key":"ref1","article-title":"Translation tutorial: 21 fairness definitions and their politics","author":"narayanan","year":"0","journal-title":"Conference on Fairness Accountability and Transparency"},{"key":"ref20","article-title":"UCI machine learning repository","author":"dheeru","year":"2017"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1016\/j.dss.2014.03.001"},{"key":"ref21","article-title":"Machine bias: There's software used across the country to predict future criminals. And it's biased against blacks","author":"angwin","year":"2016","journal-title":"ProPublica"},{"key":"ref24","article-title":"Medical Expenditure Panel survey data: 2016 full year consolidated data file","year":"2016"},{"key":"ref23","article-title":"Medical Expenditure Panel Survey data: 2015 Full Year Consolidated Data File","year":"2015"},{"key":"ref26","first-page":"3992","article-title":"Optimized pre-processing for discrimination prevention","author":"calmon","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref25","first-page":"325","article-title":"Learning fair representations","author":"zemel","year":"2013","journal-title":"Proc Int Conf Mach Learn"}],"container-title":["IBM Journal of Research and Development"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/5288520\/8895606\/08843908.pdf?arnumber=8843908","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,20]],"date-time":"2025-10-20T17:55:34Z","timestamp":1760982934000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8843908\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,7,1]]},"references-count":37,"journal-issue":{"issue":"4\/5"},"URL":"https:\/\/doi.org\/10.1147\/jrd.2019.2942287","relation":{},"ISSN":["0018-8646","0018-8646"],"issn-type":[{"value":"0018-8646","type":"print"},{"value":"0018-8646","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019,7,1]]}}}