{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T23:47:16Z","timestamp":1773359236641,"version":"3.50.1"},"reference-count":73,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2023,6,19]],"date-time":"2023-06-19T00:00:00Z","timestamp":1687132800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,6,19]],"date-time":"2023-06-19T00:00:00Z","timestamp":1687132800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100013348","name":"Innosuisse - Schweizerische Agentur f\u00fcr Innovationsf\u00f6rderung","doi-asserted-by":"publisher","award":["44692.1 IP-SBM"],"award-info":[{"award-number":["44692.1 IP-SBM"]}],"id":[{"id":"10.13039\/501100013348","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100018694","name":"HORIZON EUROPE Marie Sklodowska-Curie Actions","doi-asserted-by":"publisher","award":["898322"],"award-info":[{"award-number":["898322"]}],"id":[{"id":"10.13039\/100018694","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001711","name":"Schweizerischer Nationalfonds zur F\u00f6rderung der Wissenschaftlichen Forschung","doi-asserted-by":"publisher","award":["187473"],"award-info":[{"award-number":["187473"]}],"id":[{"id":"10.13039\/501100001711","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Philos. Technol."],"published-print":{"date-parts":[[2023,9]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Algorithmic predictions are promising for insurance companies to develop personalized risk models for determining premiums. In this context, issues of fairness, discrimination, and social injustice might arise: Algorithms for estimating the risk based on personal data may be biased towards specific social groups, leading to systematic disadvantages for those groups. Personalized premiums may thus lead to discrimination and social injustice. It is well known from many application fields that such biases occur frequently and naturally when prediction models are applied to people unless special efforts are made to avoid them. Insurance is no exception. In this paper, we provide a thorough analysis of algorithmic fairness in the case of insurance premiums. We ask what \u201cfairness\u201d might mean in this context and how the fairness of a premium system can be measured. For this, we apply the established fairness frameworks of the fair machine learning literature to the case of insurance premiums and show which of the existing fairness criteria can be applied to assess the fairness of insurance premiums. We argue that two of the often-discussed group fairness criteria, <jats:italic>independence<\/jats:italic> (also called <jats:italic>statistical parity<\/jats:italic> or <jats:italic>demographic parity<\/jats:italic>) and <jats:italic>separation<\/jats:italic> (also known as <jats:italic>equalized odds<\/jats:italic>), are not normatively appropriate for insurance premiums. Instead, we propose the <jats:italic>sufficiency<\/jats:italic> criterion (also known as <jats:italic>well-calibration<\/jats:italic>) as a morally defensible alternative that allows us to test for systematic biases in premiums towards certain groups based on the risk they bring to the pool. In addition, we clarify the connection between group fairness and different degrees of personalization. Our findings enable insurers to assess the fairness properties of their risk models, helping them avoid reputation damage resulting from potentially unfair and discriminatory premium systems.<\/jats:p>","DOI":"10.1007\/s13347-023-00624-9","type":"journal-article","created":{"date-parts":[[2023,6,19]],"date-time":"2023-06-19T17:01:34Z","timestamp":1687194094000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":21,"title":["Fairness and Risk: An Ethical Argument for a Group Fairness Definition Insurers Can Use"],"prefix":"10.1007","volume":"36","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2019-4829","authenticated-orcid":false,"given":"Joachim","family":"Baumann","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7053-4724","authenticated-orcid":false,"given":"Michele","family":"Loi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,19]]},"reference":[{"key":"624_CR1","unstructured":"Agarwal, A., Dud\u00edk, M., & Wu, Z.\u00a0S. (2019). Fair Regression: Quantitative Definitions and Reduction-based Algorithms. 36th International Conference on Machine Learning, ICML 2019, 2019-June, 166\u2013183. arXiv:1905.12843"},{"key":"624_CR2","unstructured":"Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016). Machine bias. ProPublica, May, 23(2016), 139\u2013159. https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing"},{"key":"624_CR3","unstructured":"Aristotle (1984a). Nicomachean Ethics. In J. Barnes (Ed.), Complete Works of Aristotle. Princeton University Press."},{"key":"624_CR4","unstructured":"Aristotle (1984b). Politics. In J. Barnes (Ed.), Complete Works of Aristotle. Princeton University Press."},{"key":"624_CR5","unstructured":"Arrow, K.\u00a0J. (1963). Uncertainty and the Welfare Economics of Medical Care. The American Economic Review, 53(5), 941\u2013973. http:\/\/www.jstor.org\/stable\/1812044"},{"issue":"1\u20132","key":"624_CR6","doi-asserted-by":"publisher","first-page":"131","DOI":"10.3818\/JRP.4.1.2002.131","volume":"4","author":"I Ayres","year":"2002","unstructured":"Ayres, I. (2002). Outcome Tests of Racial Disparities in Police Practices. Justice Research and Policy, 4(1\u20132), 131\u2013142. https:\/\/doi.org\/10.3818\/JRP.4.1.2002.131","journal-title":"Justice Research and Policy"},{"key":"624_CR7","unstructured":"Barocas, S., Hardt, M., & Narayanan, A. (2019). Fairness and Machine Learning. fairmlbook.org."},{"key":"624_CR8","doi-asserted-by":"crossref","unstructured":"Barocas, S. & Selbst, A.\u00a0D. (2016). Big Data\u2019s Disparate Impact. California Law Review, 104(3):671\u2013732. http:\/\/www.jstor.org\/stable\/24758720","DOI":"10.2139\/ssrn.2477899"},{"key":"624_CR9","doi-asserted-by":"publisher","unstructured":"Baumann, J., Hann\u00e1k, A., & Heitz, C. (2022). Enforcing Group Fairness in Algorithmic Decision Making: Utility Maximization Under Sufficiency. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, FAccT \u201922 (pp. 2315\u20132326). New York, NY, USA: Association for Computing Machinery. https:\/\/doi.org\/10.1145\/3531146.3534645","DOI":"10.1145\/3531146.3534645"},{"key":"624_CR10","doi-asserted-by":"publisher","unstructured":"Baumann, J. & Heitz, C. (2022). Group Fairness in Prediction-Based Decision Making: From Moral Assessment to Implementation. In 2022 9th Swiss Conference on Data Science (SDS) (pp. 19\u201325). https:\/\/doi.org\/10.1109\/SDS54800.2022.00011","DOI":"10.1109\/SDS54800.2022.00011"},{"key":"624_CR11","unstructured":"Berk, R., Heidari, H., Jabbari, S., Joseph, M., Kearns, M., Morgenstern, J., Neel, S., and Roth, A. (2017). A Convex Framework for Fair Regression. arXiv preprint arXiv:1706.02409."},{"issue":"1","key":"624_CR12","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1177\/0049124118782533","volume":"50","author":"R Berk","year":"2021","unstructured":"Berk, R., Heidari, H., Jabbari, S., Kearns, M., & Roth, A. (2021). Fairness in Criminal Justice Risk Assessments: The State of the Art. Sociological Methods & Research, 50(1), 3\u201344. https:\/\/doi.org\/10.1177\/0049124118782533","journal-title":"Sociological Methods & Research"},{"key":"624_CR13","unstructured":"Caton, S. & Haas, C. (2020). Fairness in Machine Learning: A Survey. arXiv:2010.04053"},{"issue":"2","key":"624_CR14","doi-asserted-by":"publisher","first-page":"205395172093922","DOI":"10.1177\/2053951720939228","volume":"7","author":"A Cevolini","year":"2020","unstructured":"Cevolini, A., & Esposito, E. (2020). From pool to profile: Social consequences of algorithmic prediction in insurance. Big Data & Society, 7(2), 2053951720939228. https:\/\/doi.org\/10.1177\/2053951720939228","journal-title":"Big Data & Society"},{"key":"624_CR15","doi-asserted-by":"publisher","unstructured":"Charpentier, A., (Ed.). (2014). Computational actuarial science with R. CRC press. https:\/\/doi.org\/10.1201\/b17230","DOI":"10.1201\/b17230"},{"issue":"2","key":"624_CR16","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1089\/big.2016.0047","volume":"5","author":"A Chouldechova","year":"2017","unstructured":"Chouldechova, A. (2017). Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Prediction Instruments. Big data, 5(2), 153\u2013163. https:\/\/doi.org\/10.1089\/big.2016.0047","journal-title":"Big data"},{"key":"624_CR17","unstructured":"Chouldechova, A., Benavides-Prado, D., Fialko, O., & Vaithianathan, R. (2018). A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions. In Friedler, S.\u00a0A. & Wilson, C., (Eds.), Proceedings of the 1st Conference on Fairness, Accountability and Transparency, vol.\u00a081 of Proceedings of Machine Learning Research (pp. 134\u2013148). PMLR. https:\/\/proceedings.mlr.press\/v81\/chouldechova18a.html"},{"key":"624_CR18","unstructured":"Corbett-Davies, S. & Goel, S. (2018). The Measure and Mismeasure of Fairness: A Critical Review of Fair Machine Learning. arXiv:1808.00023"},{"key":"624_CR19","doi-asserted-by":"publisher","unstructured":"Corbett-Davies, S., Pierson, E., Feller, A., Goel, S., & Huq, A. (2017). Algorithmic Decision Making and the Cost of Fairness. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD \u201917 (pp. 797\u2013806). New York, NY, USA: Association for Computing Machinery. https:\/\/doi.org\/10.1145\/3097983.3098095","DOI":"10.1145\/3097983.3098095"},{"key":"624_CR20","unstructured":"Council of the European Union (2004). Directive 2004\/113\/EC Implementing the principle of equal treatment between men and women in the access to and supply of goods and services. Official Journal of the European Union (L 373, 21.12.2004):37\u201343. https:\/\/eur-lex.europa.eu\/legal-content\/EN\/TXT\/?uri=celex%3A32004L0113"},{"key":"624_CR21","unstructured":"Daniels, N. (1981). Health-Care Needs and Distributive Justice. Philosophy & Public Affairs, 10(2), 146\u2013179. http:\/\/www.jstor.org\/stable\/2264976"},{"key":"624_CR22","unstructured":"Daniels, N. (2004). The functions of insurance and the fairness of genetic underwriting (pp. 119\u2013145). Genetics and life insurance: Medical underwriting and social policy."},{"key":"624_CR23","unstructured":"Dolman, C., Lazar, S., Caetano, T., & Semenovich, D. (2020). Should I Use That Rating Factor? A Philosophical Approach to an Old Problem. In 20\/20 All-Actuaries Virtual Summit, volume\u00a061, (pp. 0\u201323)."},{"key":"624_CR24","unstructured":"Dolman, C. & Semenovich, D. (2018). Actuarial Fairness. In Workshop on Ethical, Social and Governance Issues in AI, NIPS 2018."},{"key":"624_CR25","doi-asserted-by":"publisher","unstructured":"Donahue, K. & Barocas, S. (2021). Better Together?: How Externalities of Size Complicate Notions of Solidarity and Actuarial Fairness. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (pp. 185\u2013195). New York, NY, USA: ACM. https:\/\/doi.org\/10.1145\/3442188.3445882","DOI":"10.1145\/3442188.3445882"},{"key":"624_CR26","doi-asserted-by":"publisher","unstructured":"Dwork, C., Hardt, M., Pitassi, T., Reingold, O., & Zemel, R. (2012). Fairness through awareness. In ITCS 2012 - Innovations in Theoretical Computer Science Conference (pp. 214\u2013226). New York, New York, USA: ACM Press. https:\/\/doi.org\/10.1145\/2090236.2090255, arXiv:1104.3913","DOI":"10.1145\/2090236.2090255"},{"key":"624_CR27","unstructured":"Dworkin, R. (1981). What is Equality? Part 2: Equality of Resources. Philosophy & Public Affairs, 10(4), 283\u2013345. http:\/\/www.jstor.org\/stable\/2265047"},{"key":"624_CR28","doi-asserted-by":"publisher","DOI":"10.1515\/9781478009214","author":"F Ewald","year":"2020","unstructured":"Ewald, F., & Johnson, T. S. (2020). The Birth of Solidarity. Duke University Press. https:\/\/doi.org\/10.1515\/9781478009214","journal-title":"Duke University Press"},{"issue":"4","key":"624_CR29","doi-asserted-by":"publisher","first-page":"136","DOI":"10.1145\/3433949","volume":"64","author":"SA Friedler","year":"2021","unstructured":"Friedler, S. A., Scheidegger, C., & Venkatasubramanian, S. (2021). The (Im)possibility of fairness. Communications of the ACM, 64(4), 136\u2013143. https:\/\/doi.org\/10.1145\/3433949","journal-title":"Communications of the ACM"},{"key":"624_CR30","unstructured":"Friedman, M. (1970). A Friedman doctrine\u2013 The Social Responsibility of Business is to Increase its Profits. The New York Times. https:\/\/www.nytimes.com\/1970\/09\/13\/archives\/a-friedman-doctrine-the-social-responsibility-of-business-is-to.html"},{"key":"624_CR31","doi-asserted-by":"publisher","unstructured":"Friedman, M. (2007). The Social Responsibility of Business is to Increase Its Profits. In Zimmerli, W.\u00a0C., Holzinger, M., & Richter, K., (Eds.), Corporate Ethics and Corporate Governance, (pp. 173\u2013178). Berlin, Heidelberg: Springer Berlin Heidelberg. https:\/\/doi.org\/10.1007\/978-3-540-70818-6_14","DOI":"10.1007\/978-3-540-70818-6_14"},{"key":"624_CR32","doi-asserted-by":"publisher","unstructured":"Fuster, A., Goldsmith-Pinkham, P., Ramadorai, T., & Walther, A. (2017). Predictably Unequal? The Effects of Machine Learning on Credit Markets: SSRN. https:\/\/doi.org\/10.2139\/ssrn.3072038","DOI":"10.2139\/ssrn.3072038"},{"key":"624_CR33","doi-asserted-by":"publisher","first-page":"3662","DOI":"10.1109\/BigData50022.2020.9378025","volume-title":"2020 IEEE International Conference on Big Data (Big Data)","author":"P Garg","year":"2020","unstructured":"Garg, P., Villasenor, J., & Foggo, V. (2020). Fairness metrics: A comparative analysis. 2020 IEEE International Conference on Big Data (Big Data) (pp. 3662\u20133666). Los Alamitos, CA, USA: IEEE Computer Society."},{"key":"624_CR34","unstructured":"Hardt, M., Price, E., & Srebro, N. (2016). Equality of opportunity in supervised learning. In Advances in Neural Information Processing Systems, NIPS\u201916, (pp. 3323\u20133331). Red Hook, NY, USA: Curran Associates Inc. arXiv:1610.02413"},{"key":"624_CR35","unstructured":"Hebert-Johnson, U., Kim, M., Reingold, O., & Rothblum, G. (2018). Multicalibration: Calibration for the (Computationally-Identifiable) Masses. In Dy, J. & Krause, A., (Eds.), Proceedings of the 35th International Conference on Machine Learning, volume\u00a080 of Proceedings of Machine Learning Research, (pp. 1939\u20131948). PMLR. https:\/\/proceedings.mlr.press\/v80\/hebert-johnson18a.html"},{"issue":"2","key":"624_CR36","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1111\/papa.12189","volume":"49","author":"B Hedden","year":"2021","unstructured":"Hedden, B. (2021). On statistical criteria of algorithmic fairness. Philosophy & Public Affairs, 49(2), 209\u2013231. https:\/\/doi.org\/10.1111\/papa.12189","journal-title":"Philosophy & Public Affairs"},{"key":"624_CR37","unstructured":"Kearns, M., Neel, S., Roth, A., & Wu, Z.\u00a0S. (2018). Preventing Fairness Gerrymandering: Auditing and Learning for Subgroup Fairness. In Dy, J. & Krause, A., (Eds.), Proceedings of the 35th International Conference on Machine Learning, vol.\u00a080 of Proceedings of Machine Learning Research, (pp. 2564\u20132572). PMLR. https:\/\/proceedings.mlr.press\/v80\/kearns18a.html"},{"key":"624_CR38","volume-title":"The Ethical Algorithm: The Science of Socially Aware Algorithm Design","author":"M Kearns","year":"2019","unstructured":"Kearns, M., & Roth, A. (2019). The Ethical Algorithm: The Science of Socially Aware Algorithm Design. USA: Oxford University Press Inc."},{"key":"624_CR39","doi-asserted-by":"publisher","first-page":"22","DOI":"10.1257\/pandp.20181018","volume":"108","author":"J Kleinberg","year":"2018","unstructured":"Kleinberg, J., Ludwig, J., Mullainathan, S., & Rambachan, A. (2018). Algorithmic Fairness. AEA Papers and Proceedings, 108, 22\u201327. https:\/\/doi.org\/10.1257\/pandp.20181018","journal-title":"Algorithmic Fairness. AEA Papers and Proceedings"},{"key":"624_CR40","unstructured":"Kleinberg, J., Mullainathan, S., & Raghavan, M. (2016). Inherent Trade-Offs in the Fair Determination of Risk Scores. arXiv:1609.05807v2"},{"key":"624_CR41","unstructured":"Knight, F.\u00a0H. (1921). Risk, uncertainty and profit. Houghton Mifflin."},{"key":"624_CR42","unstructured":"Kusner, M.\u00a0J., Loftus, J., Russell, C., & Silva, R. (2017). Counterfactual Fairness. In Guyon, I., Luxburg, U.\u00a0V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., & Garnett, R., (Eds.), Advances in Neural Information Processing Systems, vol.\u00a030. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2017\/file\/a486cd07e4ac3d270571622f4f316ec5-Paper.pdf"},{"issue":"3","key":"624_CR43","doi-asserted-by":"publisher","first-page":"519","DOI":"10.1007\/s10551-014-2120-0","volume":"128","author":"X Landes","year":"2015","unstructured":"Landes, X. (2015). How Fair Is Actuarial Fairness? Journal of Business Ethics, 128(3), 519\u2013533. https:\/\/doi.org\/10.1007\/s10551-014-2120-0","journal-title":"Journal of Business Ethics"},{"issue":"1","key":"624_CR44","doi-asserted-by":"publisher","first-page":"33","DOI":"10.1007\/s10551-012-1221-x","volume":"103","author":"T-K Lehtonen","year":"2011","unstructured":"Lehtonen, T.-K., & Liukko, J. (2011). The Forms and Limits of Insurance Solidarity. Journal of Business Ethics, 103(1), 33\u201344. https:\/\/doi.org\/10.1007\/s10551-012-1221-x","journal-title":"Journal of Business Ethics"},{"issue":"1","key":"624_CR45","doi-asserted-by":"publisher","first-page":"55","DOI":"10.1017\/asb.2021.23","volume":"52","author":"M Lindholm","year":"2022","unstructured":"Lindholm, M., Richman, R., Tsanakas, A., & W\u00fcthrich, M. V. (2022). Discrimination-Free Insurance Pricing. ASTIN. Bulletin, 52(1), 55\u201389. https:\/\/doi.org\/10.1017\/asb.2021.23","journal-title":"Bulletin"},{"issue":"4","key":"624_CR46","doi-asserted-by":"publisher","first-page":"385","DOI":"10.1111\/j.1467-9760.2007.00285.x","volume":"15","author":"K Lippert-Rasmussen","year":"2007","unstructured":"Lippert-Rasmussen, K. (2007). Nothing Personal: On Statistical Discrimination*. Journal of Political Philosophy, 15(4), 385\u2013403. https:\/\/doi.org\/10.1111\/j.1467-9760.2007.00285.x","journal-title":"Journal of Political Philosophy"},{"key":"624_CR47","volume-title":"Born free and equal? a philosophical inquiry into the nature of discrimination","author":"K Lippert-Rasmussen","year":"2014","unstructured":"Lippert-Rasmussen, K. (2014). Born free and equal? a philosophical inquiry into the nature of discrimination. New York: Oxford University Press, Oxford."},{"key":"624_CR48","unstructured":"Lipton, Z.\u00a0C., Chouldechova, A., and McAuley, J. (2018). Does mitigating ML\u2019s impact disparity require treatment disparity? In Proceedings of the 32nd International Conference on Neural Information Processing Systems, (pp. 8136\u20138146). Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2018\/file\/8e0384779e58ce2af40eb365b318cc32-Paper.pdf"},{"key":"624_CR49","doi-asserted-by":"publisher","DOI":"10.1007\/s13347-021-00444-9","author":"M Loi","year":"2021","unstructured":"Loi, M., & Christen, M. (2021). Choosing how to discriminate: navigating ethical trade-offs in fair algorithmic design for the insurance sector. Philosophy & Technology. https:\/\/doi.org\/10.1007\/s13347-021-00444-9","journal-title":"Philosophy & Technology"},{"key":"624_CR50","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.3450300","author":"M Loi","year":"2019","unstructured":"Loi, M., Herlitz, A., & Heidari, H. (2019). A Philosophical Theory of Fairness for Prediction-Based Decisions. SSRN Electronic Journal. https:\/\/doi.org\/10.2139\/ssrn.3450300","journal-title":"SSRN Electronic Journal"},{"key":"624_CR51","doi-asserted-by":"crossref","unstructured":"Lorentzen, C. and Mayer, M. (2020). Peeking into the Black Box: An Actuarial Case Study for Interpretable Machine Learning. SSRN. https:\/\/ssrn.com\/abstract=3595944","DOI":"10.2139\/ssrn.3595944"},{"key":"624_CR52","unstructured":"Miller, M.\u00a0J. (2009). Disparate Impact and Unfairly Discriminatory Insurance Rates. In Casualty Actuarial Society E-Forum, Winter 2009, (pp. 276\u2013288). Citeseer. https:\/\/www.casact.org\/sites\/default\/files\/database\/forum_09wforum_miller.pdf"},{"key":"624_CR53","volume-title":"Translation tutorial: 21 fairness definitions and their politics","author":"A Narayanan","year":"2018","unstructured":"Narayanan, A. (2018). Translation tutorial: 21 fairness definitions and their politics. New York, USA: In Proc. Conf. Fairness Accountability Transp."},{"key":"624_CR54","unstructured":"Noll, A., Salzmann, R., & Wuthrich, M.\u00a0V. (2020). Case study: French motor third-party liability claims. SSRN. https:\/\/ssrn.com\/abstract=3164764"},{"key":"624_CR55","doi-asserted-by":"crossref","unstructured":"Ohlsson, E. & Johansson, B. (2010). Non-life insurance pricing with generalized linear models, vol. 174. Springer.","DOI":"10.1007\/978-3-642-10791-7"},{"issue":"4","key":"624_CR56","doi-asserted-by":"publisher","first-page":"567","DOI":"10.5840\/monist20068948","volume":"89","author":"M O\u2019Neill","year":"2006","unstructured":"O\u2019Neill, M. (2006). Genetic Information, Life Insurance, and Social Justice. The Monist, 89(4), 567\u2013592. https:\/\/doi.org\/10.5840\/monist20068948","journal-title":"The Monist"},{"key":"624_CR57","unstructured":"Pessach, D. and Shmueli, E. (2020). Algorithmic fairness. arXiv preprint. arXiv:2001.09784"},{"key":"624_CR58","doi-asserted-by":"publisher","unstructured":"Potash, E., Brew, J., Loewi, A., Majumdar, S., Reece, A., Walsh, J., Rozier, E., Jorgenson, E., Mansour, R., & Ghani, R. (2015). Predictive Modeling for Public Health: Preventing Childhood Lead Poisoning. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD \u201915, (pp. 2039\u20132047), New York, NY, USA: Association for Computing Machinery. https:\/\/doi.org\/10.1145\/2783258.2788629","DOI":"10.1145\/2783258.2788629"},{"issue":"1","key":"624_CR59","doi-asserted-by":"publisher","first-page":"219","DOI":"10.1257\/jep.15.1.219","volume":"15","author":"M Rabin","year":"2001","unstructured":"Rabin, M., & Thaler, R. H. (2001). Anomalies: Risk Aversion. Journal of Economic Perspectives, 15(1), 219\u2013232. https:\/\/doi.org\/10.1257\/jep.15.1.219","journal-title":"Journal of Economic Perspectives"},{"key":"624_CR60","doi-asserted-by":"publisher","unstructured":"Raghavan, M., Barocas, S., Kleinberg, J., & Levy, K. (2020). Mitigating Bias in Algorithmic Hiring: Evaluating Claims and Practices. FAT* 2020 - Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, 469\u2013481. https:\/\/doi.org\/10.1145\/3351095.3372828","DOI":"10.1145\/3351095.3372828"},{"key":"624_CR61","doi-asserted-by":"crossref","unstructured":"Rebert, L. & Van Hoyweghen, I. (2015). The right to underwrite gender. The Goods & Services Directive and the politics of insurance pricing. Tijdschrift voor Genderstudies, 18(4):413\u2013431.","DOI":"10.5117\/TVGN2015.4.REBE"},{"key":"624_CR62","unstructured":"Reichenbach, H. (1971). The theory of probability. Univ of California Press."},{"issue":"2","key":"624_CR63","doi-asserted-by":"publisher","first-page":"423","DOI":"10.1017\/s2071832200001863","volume":"14","author":"E Schanze","year":"2013","unstructured":"Schanze, E. (2013). Injustice by Generalization: Notes on the Test-Achats Decision of the European Court of Justice. German Law Journal, 14(2), 423\u2013433. https:\/\/doi.org\/10.1017\/s2071832200001863","journal-title":"German Law Journal"},{"issue":"3","key":"624_CR64","doi-asserted-by":"publisher","first-page":"1193","DOI":"10.1214\/17-AOAS1058","volume":"11","author":"C Simoiu","year":"2017","unstructured":"Simoiu, C., Corbett-Davies, S., & Goel, S. (2017). The problem of infra-marginality in outcome tests for discrimination. The Annals of Applied Statistics, 11(3), 1193\u20131216. https:\/\/doi.org\/10.1214\/17-AOAS1058","journal-title":"The Annals of Applied Statistics"},{"key":"624_CR65","unstructured":"Steinberg, D., Reid, A., & O\u2019Callaghan, S. (2020a). Fairness Measures for Regression via Probabilistic Classification. arXiv:2001.06089"},{"key":"624_CR66","unstructured":"Steinberg, D., Reid, A., O\u2019Callaghan, S., Lattimore, F., McCalman, L., & Caetano, T. (2020b). Fast Fair Regression via Efficient Approximations of Mutual Information. arXiv:2002.06200"},{"issue":"2","key":"624_CR67","doi-asserted-by":"publisher","first-page":"190","DOI":"10.1057\/palgrave.gpp.2510078","volume":"31","author":"Y Thiery","year":"2006","unstructured":"Thiery, Y., & Van Schoubroeck, C. (2006). Fairness and Equality in Insurance Classification. The Geneva Papers on Risk and Insurance - Issues and Practice, 31(2), 190\u2013211. https:\/\/doi.org\/10.1057\/palgrave.gpp.2510078","journal-title":"The Geneva Papers on Risk and Insurance - Issues and Practice"},{"key":"624_CR68","doi-asserted-by":"publisher","unstructured":"Verma, S. & Rubin, J. (2018). Fairness Definitions Explained. In Proceedings of the International Workshop on Software Fairness, FairWare \u201918, (pp. 1\u20137). New York, NY, USA: Association for Computing Machinery. https:\/\/doi.org\/10.1145\/3194770.3194776","DOI":"10.1145\/3194770.3194776"},{"key":"624_CR69","doi-asserted-by":"crossref","unstructured":"Wakker, P.\u00a0P. (2010). Prospect theory: For risk and ambiguity. Cambridge university press.","DOI":"10.1017\/CBO9780511779329"},{"key":"624_CR70","unstructured":"Wuthrich, M.\u00a0V. (2020). Non-life insurance: mathematics & statistics. SSRN. https:\/\/ssrn.com\/abstract=2319328"},{"key":"624_CR71","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-12409-9","author":"MV W\u00fcthrich","year":"2023","unstructured":"W\u00fcthrich, M. V., & Merz, M. (2023). Statistical Foundations of Actuarial Learning and its Applications. Springer Actuarial. Springer International Publishing: Cham.https:\/\/doi.org\/10.1007\/978-3-031-12409-9","journal-title":"Springer Actuarial. Springer International Publishing: Cham."},{"key":"624_CR72","doi-asserted-by":"publisher","unstructured":"Zafar, M.\u00a0B., Valera, I., Gomez Rodriguez, M., & Gummadi, K.\u00a0P. (2017). Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment. In Proceedings of the 26th International Conference on World Wide Web, WWW \u201917, (pp. 1171\u20131180). Republic and Canton of Geneva, CHE. International World Wide Web Conferences Steering Committee. https:\/\/doi.org\/10.1145\/3038912.3052660","DOI":"10.1145\/3038912.3052660"},{"key":"624_CR73","unstructured":"Zliobaite, I. (2015). On the relation between accuracy and fairness in binary classification. arXiv:1505.05723"}],"container-title":["Philosophy &amp; Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13347-023-00624-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13347-023-00624-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13347-023-00624-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,9,20]],"date-time":"2023-09-20T11:21:50Z","timestamp":1695208910000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13347-023-00624-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,19]]},"references-count":73,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2023,9]]}},"alternative-id":["624"],"URL":"https:\/\/doi.org\/10.1007\/s13347-023-00624-9","relation":{},"ISSN":["2210-5433","2210-5441"],"issn-type":[{"value":"2210-5433","type":"print"},{"value":"2210-5441","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,6,19]]},"assertion":[{"value":"12 October 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 March 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 June 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This is a theoretical paper without any experiments involving human subjects. Thus, no ethics approval and no consent to participate is required.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval and consent to participate"}},{"value":"This is a theoretical paper without any experiments involving human subjects. Thus, no consent for publication is required.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"The authors declare no competing interests.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"45"}}