{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T14:35:22Z","timestamp":1776090922515,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":63,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T00:00:00Z","timestamp":1686528000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Innosuisse - Schweizerische Agentur f\u00fcr Innovationsf\u00f6rderung","award":["44692.1 IP-SBM"],"award-info":[{"award-number":["44692.1 IP-SBM"]}]},{"name":"Schweizerischer Nationalfonds zur F\u00f6rderung der Wissenschaftlichen Forschung","award":["187473"],"award-info":[{"award-number":["187473"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,6,12]]},"DOI":"10.1145\/3593013.3594058","type":"proceedings-article","created":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T14:40:46Z","timestamp":1686580846000},"page":"1002-1013","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":11,"title":["Bias on Demand: A Modelling Framework That Generates Synthetic Data With Bias"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2019-4829","authenticated-orcid":false,"given":"Joachim","family":"Baumann","sequence":"first","affiliation":[{"name":"Department of Informatics, University of Zurich, Switzerland and Zurich University of Applied Sciences, Switzerland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5234-1155","authenticated-orcid":false,"given":"Alessandro","family":"Castelnovo","sequence":"additional","affiliation":[{"name":"Data Science &amp; Artificial Intelligence, Intesa Sanpaolo, Italy and Dept. of Informatics, Systems and Communication, University Milano Bicocca, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-6714-5161","authenticated-orcid":false,"given":"Riccardo","family":"Crupi","sequence":"additional","affiliation":[{"name":"Data Science &amp; Artificial Intelligence, Intesa Sanpaolo, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-0048-7455","authenticated-orcid":false,"given":"Nicole","family":"Inverardi","sequence":"additional","affiliation":[{"name":"Data Science &amp; Artificial Intelligence, Intesa Sanpaolo, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2711-8343","authenticated-orcid":false,"given":"Daniele","family":"Regoli","sequence":"additional","affiliation":[{"name":"Data Science &amp; Artificial Intelligence, Intesa Sanpaolo, Italy"}]}],"member":"320","published-online":{"date-parts":[[2023,6,12]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"International Conference on Machine Learning. PMLR, 60\u201369","author":"Agarwal Alekh","year":"2018","unstructured":"Alekh Agarwal, Alina Beygelzimer, Miroslav Dud\u00edk, John Langford, and Hanna Wallach. 2018. A reductions approach to fair classification. In International Conference on Machine Learning. PMLR, 60\u201369."},{"key":"e_1_3_2_2_2_1","volume-title":"Machine bias: There\u2019s software used across the country to predict future criminals, and it\u2019s biased against blacks. ProPublica","author":"Angwin Julia","year":"2016","unstructured":"Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. 2016. Machine bias: There\u2019s software used across the country to predict future criminals, and it\u2019s biased against blacks. ProPublica (2016)."},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/3383455.3422554"},{"key":"e_1_3_2_2_4_1","unstructured":"Solon Barocas Moritz Hardt and Arvind Narayanan. 2019. Fairness and Machine Learning. fairmlbook.org. http:\/\/www.fairmlbook.org."},{"key":"e_1_3_2_2_5_1","first-page":"671","article-title":"Big data\u2019s disparate impact","volume":"104","author":"Barocas Solon","year":"2016","unstructured":"Solon Barocas and Andrew D Selbst. 2016. Big data\u2019s disparate impact. Calif. L. Rev. 104 (2016), 671.","journal-title":"Calif. L. Rev."},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3534645"},{"key":"e_1_3_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/SDS54800.2022.00011"},{"key":"e_1_3_2_2_8_1","unstructured":"Joachim Baumann Corinna Hertweck Michele Loi and Christoph Heitz. 2023. Distributive Justice as the Foundational Premise of Fair ML: Unification Extension and Interpretation of Group Fairness Metrics. (2023). arxiv:2206.02897http:\/\/arxiv.org\/abs\/2206.02897"},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.1007\/s13347-023-00624-9"},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11749-016-0481-7"},{"key":"e_1_3_2_2_11_1","volume-title":"Proceedings of the 1st Conference on Fairness, Accountability and Transparency(Proceedings of Machine Learning Research","volume":"159","author":"Binns Reuben","year":"2018","unstructured":"Reuben Binns. 2018. Fairness in Machine Learning: Lessons from Political Philosophy. In Proceedings of the 1st Conference on Fairness, Accountability and Transparency(Proceedings of Machine Learning Research, Vol. 81), Sorelle A. Friedler and Christo Wilson (Eds.). PMLR, 149\u2013159. https:\/\/proceedings.mlr.press\/v81\/binns18a.html"},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372864"},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2022.103099"},{"key":"e_1_3_2_2_14_1","volume-title":"BeFair: Addressing Fairness in the Banking Sector. In 2020 IEEE International Conference on Big Data (Big Data). IEEE, 3652\u20133661","author":"Castelnovo Alessandro","year":"2020","unstructured":"Alessandro Castelnovo, Riccardo Crupi, Giulia Del Gamba, Greta Greco, Aisha Naseer, Daniele Regoli, and Beatriz San Miguel Gonzalez. 2020. BeFair: Addressing Fairness in the Banking Sector. In 2020 IEEE International Conference on Big Data (Big Data). IEEE, 3652\u20133661."},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-022-07939-1"},{"key":"e_1_3_2_2_16_1","volume-title":"Towards Fairness Through Time. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, 647\u2013663","author":"Castelnovo Alessandro","year":"2021","unstructured":"Alessandro Castelnovo, Lorenzo Malandri, Fabio Mercorio, Mario Mezzanzanica, and Andrea Cosentini. 2021. Towards Fairness Through Time. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, 647\u2013663."},{"key":"e_1_3_2_2_17_1","volume-title":"Fairness in machine learning: A survey. arXiv preprint arXiv:2010.04053","author":"Caton Simon","year":"2020","unstructured":"Simon Caton and Christian Haas. 2020. Fairness in machine learning: A survey. arXiv preprint arXiv:2010.04053 (2020)."},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.5555\/3327144.3327272"},{"key":"e_1_3_2_2_19_1","volume-title":"Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2","author":"Chouldechova Alexandra","year":"2017","unstructured":"Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2 (2017), 153\u2013163."},{"key":"e_1_3_2_2_20_1","unstructured":"The European Commission. 2021. Proposal for a Regulation of the European Parliament and of the Council laying down harmonised rules on Artificial Intelligence (Artificial Intelligence Act) and amending certain Union legislative acts. https:\/\/digital-strategy.ec.europa.eu\/en\/library\/proposal-regulation-laying-down-harmonised-rules-artificial-intelligence."},{"key":"e_1_3_2_2_21_1","unstructured":"Sam Corbett-Davies and Sharad Goel. 2018. The Measure and Mismeasure of Fairness: A Critical Review of Fair Machine Learning. arxiv:1808.00023 [cs.CY] https:\/\/arxiv.org\/abs\/1808.00023"},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3097983.3098095"},{"key":"e_1_3_2_2_23_1","volume-title":"International conference on machine learning. PMLR, 1436\u20131445","author":"Creager Elliot","year":"2019","unstructured":"Elliot Creager, David Madras, J\u00f6rn-Henrik Jacobsen, Marissa Weis, Kevin Swersky, Toniann Pitassi, and Richard Zemel. 2019. Flexibly fair representation learning by disentanglement. In International conference on machine learning. PMLR, 1436\u20131445."},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372878"},{"key":"e_1_3_2_2_25_1","unstructured":"Media Department for Digital Culture and Sport. 2021. Establishing a pro-innovation approach to regulating AI An overview of the UK\u2019s emerging approach. https:\/\/www.gov.uk\/government\/publications\/establishing-a-pro-innovation-approach-to-regulating-ai."},{"key":"e_1_3_2_2_26_1","volume-title":"Retiring Adult: New Datasets for Fair Machine Learning. Advances in Neural Information Processing Systems 34","author":"Ding Frances","year":"2021","unstructured":"Frances Ding, Moritz Hardt, John Miller, and Ludwig Schmidt. 2021. Retiring Adult: New Datasets for Fair Machine Learning. Advances in Neural Information Processing Systems 34 (2021)."},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/2090236.2090255"},{"key":"e_1_3_2_2_28_1","unstructured":"Khaled El Emam Lucy Mosquera and Richard Hoptroff. 2020. Practical synthetic data generation: balancing privacy and the broad availability of data. O\u2019Reilly Media."},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10618-022-00854-z"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3433949"},{"key":"e_1_3_2_2_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287589"},{"key":"e_1_3_2_2_32_1","unstructured":"Moritz Hardt Eric Price and Nati Srebro. 2016. Equality of opportunity in supervised learning. In Advances in neural information processing systems. 3315\u20133323."},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287584"},{"key":"e_1_3_2_2_34_1","unstructured":"Corinna Hertweck Joachim Baumann Michele Loi Eleonora Vigan\u00f2 and Christoph Heitz. 2023. A Justice-Based Framework for the Analysis of Algorithmic Fairness-Utility Trade-Offs. (2023). arxiv:2206.02891http:\/\/arxiv.org\/abs\/2206.02891"},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445936"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/IC4.2009.4909197"},{"key":"e_1_3_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10115-011-0463-8"},{"key":"e_1_3_2_2_38_1","volume-title":"Machine Learning and Knowledge Discovery in Databases","author":"Kamishima Toshihiro","unstructured":"Toshihiro Kamishima, Shotaro Akaho, Hideki Asoh, and Jun Sakuma. 2012. Fairness-Aware Classifier with Prejudice Remover Regularizer. In Machine Learning and Knowledge Discovery in Databases, Peter A Flach, Tijl De Bie, and Nello Cristianini (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 35\u201350."},{"key":"e_1_3_2_2_39_1","volume-title":"Inherent trade-offs in the fair determination of risk scores. arXiv preprint arXiv:1609.05807","author":"Kleinberg Jon","year":"2016","unstructured":"Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan. 2016. Inherent trade-offs in the fair determination of risk scores. arXiv preprint arXiv:1609.05807 (2016)."},{"key":"e_1_3_2_2_40_1","volume-title":"A survey on datasets for fairness-aware machine learning","author":"Quy Tai Le","year":"2022","unstructured":"Tai Le Quy, Arjun Roy, Vasileios Iosifidis, Wenbin Zhang, and Eirini Ntoutsi. 2022. A survey on datasets for fairness-aware machine learning. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery (2022), e1452."},{"key":"e_1_3_2_2_41_1","volume-title":"Advances in Neural Information Processing Systems","volume":"31","author":"Lipton Zachary C.","year":"2018","unstructured":"Zachary C. Lipton, Alexandra Chouldechova, and Julian McAuley. 2018. Does mitigating ML\u2019s impact disparity require treatment disparity?. In Advances in Neural Information Processing Systems, Vol. 31. Curran Associates, Inc., 8136\u20138146. arxiv:1711.07076https:\/\/proceedings.neurips.cc\/paper\/2018\/file\/8e0384779e58ce2af40eb365b318cc32-Paper.pdf"},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1002\/widm.1326"},{"key":"e_1_3_2_2_43_1","volume-title":"Manish Bhide, Diptikalyan Saha, Kush R Varshney, and Ruchir Puri.","author":"Lohia Pranay K","year":"2019","unstructured":"Pranay K Lohia, Karthikeyan Natesan Ramamurthy, Manish Bhide, Diptikalyan Saha, Kush R Varshney, and Ruchir Puri. 2019. Bias mitigation post-processing for individual and group fairness. In Icassp 2019-2019 ieee international conference on acoustics, speech and signal processing (icassp). IEEE, 2847\u20132851."},{"key":"e_1_3_2_2_44_1","volume-title":"A Philosophical Theory of Fairness for Prediction-Based Decisions. Available at SSRN 3450300","author":"Loi Michele","year":"2019","unstructured":"Michele Loi, Anders Herlitz, and Hoda Heidari. 2019. A Philosophical Theory of Fairness for Prediction-Based Decisions. Available at SSRN 3450300 (2019)."},{"key":"e_1_3_2_2_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376445"},{"key":"e_1_3_2_2_46_1","volume-title":"Anonymization techniques for privacy preserving data publishing: A comprehensive survey","author":"Majeed Abdul","year":"2020","unstructured":"Abdul Majeed and Sungchang Lee. 2020. Anonymization techniques for privacy preserving data publishing: A comprehensive survey. IEEE access 9 (2020), 8512\u20138545."},{"key":"e_1_3_2_2_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3457607"},{"key":"e_1_3_2_2_48_1","volume-title":"The US Algorithmic Accountability Act of 2022 vs. The EU Artificial Intelligence Act: what can they learn from each other?Minds and Machines","author":"M\u00f6kander Jakob","year":"2022","unstructured":"Jakob M\u00f6kander, Prathm Juneja, David S Watson, and Luciano Floridi. 2022. The US Algorithmic Accountability Act of 2022 vs. The EU Artificial Intelligence Act: what can they learn from each other?Minds and Machines (2022), 1\u20138."},{"key":"e_1_3_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1002\/widm.1356"},{"key":"e_1_3_2_2_50_1","unstructured":"White House Office of Science and Technology Policy. 2022. Blueprint for an AI Bill of Rights: Making Automated Systems Work For The American People. https:\/\/www.whitehouse.gov\/ostp\/ai-bill-of-rights\/."},{"key":"e_1_3_2_2_51_1","volume-title":"Saverio Bolognani, and Anik\u00f3 Hann\u00e1k.","author":"Pagan Nicol\u00f2","year":"2023","unstructured":"Nicol\u00f2 Pagan, Joachim Baumann, Ezzat Elokda, Giulia De Pasquale, Saverio Bolognani, and Anik\u00f3 Hann\u00e1k. 2023. A Classification of Feedback Loops and Their Relation to Biases in Automated Decision-Making Systems. (2023). arxiv:2305.06055http:\/\/arxiv.org\/abs\/2305.06055"},{"key":"e_1_3_2_2_52_1","unstructured":"Judea Pearl. 2009. Causality. Cambridge university press."},{"key":"e_1_3_2_2_53_1","volume-title":"The book of why: the new science of cause and effect","author":"Pearl Judea","unstructured":"Judea Pearl and Dana Mackenzie. 2018. The book of why: the new science of cause and effect. Basic Books."},{"key":"e_1_3_2_2_54_1","article-title":"Causal discovery with continuous additive noise models","volume":"15","author":"Peters Jonas","year":"2014","unstructured":"Jonas Peters, Joris M Mooij, Dominik Janzing, and Bernhard Sch\u00f6lkopf. 2014. Causal discovery with continuous additive noise models. Journal of Machine Learning Research 15, 58 (2014).","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_2_55_1","doi-asserted-by":"publisher","DOI":"10.1146\/annurev-statistics-040720-031848"},{"key":"e_1_3_2_2_56_1","volume-title":"Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1).","author":"Reddy Charan","year":"2021","unstructured":"Charan Reddy, Deepak Sharma, Soroush Mehri, Adriana Romero-Soriano, Samira Shabanian, and Sina Honari. 2021. Benchmarking bias mitigation algorithms in representation learning through fairness metrics. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1)."},{"key":"e_1_3_2_2_57_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-021-00396-x"},{"key":"e_1_3_2_2_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306618.3314314"},{"key":"e_1_3_2_2_59_1","doi-asserted-by":"crossref","unstructured":"Harini Suresh and John Guttag. 2021. A framework for understanding sources of harm throughout the machine learning life cycle. In Equity and access in algorithms mechanisms and optimization. 1\u20139.","DOI":"10.1145\/3465416.3483305"},{"key":"e_1_3_2_2_60_1","volume-title":"Multicultural experiences reduce intergroup bias through epistemic unfreezing.Journal of personality and social psychology 103, 5","author":"Tadmor Carmit T","year":"2012","unstructured":"Carmit T Tadmor, Ying-yi Hong, Melody M Chao, Fon Wiruchnipawan, and Wei Wang. 2012. Multicultural experiences reduce intergroup bias through epistemic unfreezing.Journal of personality and social psychology 103, 5 (2012), 750."},{"key":"e_1_3_2_2_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3194770.3194776"},{"key":"e_1_3_2_2_62_1","doi-asserted-by":"publisher","DOI":"10.5555\/3322706.3362016"},{"key":"e_1_3_2_2_63_1","volume-title":"Manuel Gomez Rogriguez, and Krishna P Gummadi","author":"Zafar Muhammad Bilal","year":"2017","unstructured":"Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rogriguez, and Krishna P Gummadi. 2017. Fairness constraints: Mechanisms for fair classification. In Artificial intelligence and statistics. PMLR, 962\u2013970."}],"event":{"name":"FAccT '23: the 2023 ACM Conference on Fairness, Accountability, and Transparency","location":"Chicago IL USA","acronym":"FAccT '23"},"container-title":["2023 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594058","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3593013.3594058","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:18Z","timestamp":1750178238000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594058"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,12]]},"references-count":63,"alternative-id":["10.1145\/3593013.3594058","10.1145\/3593013"],"URL":"https:\/\/doi.org\/10.1145\/3593013.3594058","relation":{},"subject":[],"published":{"date-parts":[[2023,6,12]]},"assertion":[{"value":"2023-06-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}