{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T18:22:15Z","timestamp":1776277335136,"version":"3.50.1"},"reference-count":64,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2023,5,9]],"date-time":"2023-05-09T00:00:00Z","timestamp":1683590400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,5,9]],"date-time":"2023-05-09T00:00:00Z","timestamp":1683590400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001868","name":"National Science Council","doi-asserted-by":"publisher","award":["MOST110CD804-1"],"award-info":[{"award-number":["MOST110CD804-1"]}],"id":[{"id":"10.13039\/501100001868","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["AI &amp; Soc"],"published-print":{"date-parts":[[2024,8]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Across the world, artificial intelligence (AI) technologies are being more widely employed in public sector decision-making and processes as a supposedly neutral and an efficient method for optimizing delivery of services. However, the deployment of these technologies has also prompted investigation into the potentially unanticipated consequences of their introduction, to both positive and negative ends. This paper chooses to focus specifically on the relationship between gender bias and AI, exploring claims of the neutrality of such technologies and how its understanding of bias could influence policy and outcomes. Building on a rich seam of literature from both technological and sociological fields, this article constructs an original framework through which to analyse both the perpetuation and mitigation of gender biases, choosing to categorize AI technologies based on whether their input is text or images. Through the close analysis and pairing of four case studies, the paper thus unites two often disparate approaches to the investigation of bias in technology, revealing the large and varied potential for AI to echo and even amplify existing human bias, while acknowledging the important role AI itself can play in reducing or reversing these effects. The conclusion calls for further collaboration between scholars from the worlds of technology, gender studies and public policy in fully exploring algorithmic accountability as well as in accurately and transparently exploring the potential consequences of the introduction of AI technologies.<\/jats:p>","DOI":"10.1007\/s00146-023-01675-4","type":"journal-article","created":{"date-parts":[[2023,5,11]],"date-time":"2023-05-11T01:21:44Z","timestamp":1683768104000},"page":"2045-2057","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":107,"title":["Gender bias\u00a0perpetuation and mitigation in AI technologies: challenges and opportunities"],"prefix":"10.1007","volume":"39","author":[{"given":"Sinead","family":"O\u2019Connor","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1968-2171","authenticated-orcid":false,"given":"Helen","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,5,9]]},"reference":[{"issue":"6","key":"1675_CR1","doi-asserted-by":"publisher","first-page":"917","DOI":"10.1111\/puar.12979","volume":"78","author":"PK Agarwal","year":"2018","unstructured":"Agarwal PK (2018) Public administration challenges in the world of AI and bots. Public Adm Rev 78(6):917\u2013921","journal-title":"Public Adm Rev"},{"key":"1675_CR2","unstructured":"Alizadeh K (2021) Word Vectors and Word Meanings. Medium. https:\/\/towardsdatascience.com\/word-vectors-and-word-meaning-90493d13af76. Accessed 18 Jan 2023"},{"issue":"1","key":"1675_CR3","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1093\/jopart\/muac007","volume":"33","author":"S Alon-Barkat","year":"2022","unstructured":"Alon-Barkat S, Busuioc M (2022) Human-AI interactions in public sector decision-making: \u2018Automation Bias\u2019 and \u2018Selective Adherence\u2019 to Algorithmic Advice. J Pub Adm Res Theory 33(1):153\u2013169","journal-title":"J Pub Adm Res Theory"},{"issue":"2","key":"1675_CR4","doi-asserted-by":"publisher","first-page":"134","DOI":"10.5539\/elt.v5n2p134","volume":"5","author":"M Amini","year":"2012","unstructured":"Amini M, Birjandi P (2012) Gender bias in the Iranian High School EFL Textbooks. Engl Lang Teach 5(2):134\u2013147","journal-title":"Engl Lang Teach"},{"issue":"4\/5","key":"1675_CR5","doi-asserted-by":"publisher","first-page":"4:1","DOI":"10.1147\/JRD.2019.2942287","volume":"63","author":"RK Bellamy","year":"2019","unstructured":"Bellamy RK, Dey K, Hind M et al (2019) AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias. IBM J Res Dev 63(4\/5):4:1-4:15","journal-title":"IBM J Res Dev"},{"issue":"5","key":"1675_CR6","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1109\/MIC.2021.3097604","volume":"25","author":"M Bernagozzi","year":"2021","unstructured":"Bernagozzi M, Srivastava B, Rossi F, Usmani S (2021) Gender bias in online language translators: visualization, human perception, and bias\/accuracy tradeoffs. IEEE Internet Comput 25(5):53\u201363","journal-title":"IEEE Internet Comput"},{"key":"1675_CR7","doi-asserted-by":"publisher","unstructured":"Blodgett SL, Barocas S, Daum\u00e9 III H, Wallach H (2020) Language (Technology) is Power: A Critical Survey of \u201cBias\u201d in NLP. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.485","DOI":"10.18653\/v1\/2020.acl-main.485"},{"key":"1675_CR8","first-page":"4349","volume":"29","author":"T Bolukbasi","year":"2016","unstructured":"Bolukbasi T, Chang KW, Zou JY, Saligrama V, Kalai AT (2016) Man is to computer programmer as woman is to homemaker? debiasing word embeddings. Adv Neural Inf Process Syst 29:4349\u20134357","journal-title":"Adv Neural Inf Process Syst"},{"key":"1675_CR9","unstructured":"Bryson J, Etlinger S, Keyes O, Rankin JL (2020) Gender Bias in Technology: How Far Have We Come and What Comes Next? CIGI. https:\/\/www.cigionline.org\/articles\/gender-bias-technology-how-far-have-we-come-and-what-comes-next\/?utm_medium=social&utm_source=twitter. Accessed 17 Jan 2023"},{"key":"1675_CR10","first-page":"77","volume":"81","author":"J Buolamwini","year":"2018","unstructured":"Buolamwini J, Gebru T (2018) Gender shades: Intersectional accuracy disparities in commercial gender classification. Conf Fairness Account Transpar 81:77\u201391","journal-title":"Conf Fairness Account Transpar"},{"issue":"5","key":"1675_CR11","doi-asserted-by":"publisher","first-page":"825","DOI":"10.1111\/puar.13293","volume":"81","author":"M Busuioc","year":"2021","unstructured":"Busuioc M (2021) Accountable artificial intelligence: Holding algorithms to account. Public Adm Rev 81(5):825\u2013836","journal-title":"Public Adm Rev"},{"key":"1675_CR13","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-04468-8","volume-title":"An introduction to data","author":"F Corea","year":"2019","unstructured":"Corea F (2019) An introduction to data. Springer, Cham"},{"key":"1675_CR14","unstructured":"Council of Europe (2023) Sex and gender. https:\/\/www.coe.int\/en\/web\/gender-matters\/sex-and-gender. Accessed 17 Jan 2023"},{"key":"1675_CR15","unstructured":"Crenshaw K (1989) Demarginalizing the Intersection of Race and Sex: A Black Feminist Critique of Antidiscrimination Doctrine, Feminist Theory and Antiracist Politics. University of Chicago Legal Forum, 1989(1), Article 8"},{"key":"1675_CR16","doi-asserted-by":"crossref","unstructured":"Dastin J (2022) Amazon Scraps Secret AI Recruiting Tool that Showed Bias against Women. In Martin K (ed) Ethics of Data and Analytics, 1st edn. Auerbach Publications, pp. 296\u2013299","DOI":"10.1201\/9781003278290-44"},{"key":"1675_CR17","unstructured":"Datagen (2022) MS COCO Dataset: Using it in Your Computer Vision Projects. https:\/\/datagen.tech\/guides\/image-datasets\/ms-coco-dataset-using-it-in-your-computer-vision-projects\/. Accessed 17 Jan 2023"},{"key":"1675_CR18","doi-asserted-by":"publisher","unstructured":"Deshpande KV, Pan S, Foulds JR (2020 July). Mitigating demographic Bias in AI-based resume filtering. In Adjunct Publication of the 28th ACM Conference on User Modeling, Adaptation and Personalization (pp. 268\u2013275). https:\/\/doi.org\/10.1145\/3386392.3399569","DOI":"10.1145\/3386392.3399569"},{"key":"1675_CR19","unstructured":"Dhar P, Gleason J, Souri H, Castillo CD, Chellappa R (2020) Towards gender-neutral face descriptors for mitigating bias in face recognition. arXiv preprint https:\/\/arxiv.org\/abs\/2006.07845"},{"issue":"4","key":"1675_CR20","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1145\/3512728","volume":"65","author":"T Di Noia","year":"2022","unstructured":"Di Noia T, Tintarev N, Fatourou P, Schedl M (2022) Recommender systems under European AI regulations. Commun ACM 65(4):69\u201373","journal-title":"Commun ACM"},{"key":"1675_CR21","unstructured":"Di\u00f1o G (2019) He Said, She Said: Addressing Gender in Neural Machine Translation. Slator. https:\/\/slator.com\/he-said-she-said-addressing-gender-in-neural-machine-translation\/. Accessed 17 Jan 2023"},{"key":"1675_CR22","unstructured":"Domnich A, Anbarjafari G (2021) Responsible AI: Gender bias assessment in emotion recognition. arXiv preprint arXiv:2103.11436"},{"key":"1675_CR23","doi-asserted-by":"publisher","unstructured":"Ethics guidelines for trustworthy AI. Publications Office. https:\/\/data.europa.eu\/doi\/https:\/\/doi.org\/10.2759\/177365. Accessed 17 Jan 2023","DOI":"10.2759\/177365"},{"key":"1675_CR12","doi-asserted-by":"publisher","unstructured":"European Commission, Directorate-General for Communications Networks, Content and Technology (2019) Ethics guidelines for trustworthy AI, Publications Office. https:\/\/data.europa.eu\/https:\/\/doi.org\/10.2759\/346720. Accessed 23 April 2023","DOI":"10.2759\/346720"},{"key":"1675_CR24","unstructured":"European Institute for Gender Equality (2023) Gender Bias. https:\/\/eige.europa.eu\/thesaurus\/terms\/1155. Accessed 3rd Feb 2023"},{"issue":"2","key":"1675_CR25","doi-asserted-by":"publisher","first-page":"115","DOI":"10.3233\/IP-200303","volume":"26","author":"MK Feeney","year":"2021","unstructured":"Feeney MK, Fusi F (2021) A critical analysis of the study of gender and technology in government. Inform Polity 26(2):115\u2013129","journal-title":"Inform Polity"},{"key":"1675_CR26","unstructured":"Feldman T, Peake A (2021) End-To-End Bias Mitigation: Removing Gender Bias in Deep Learning. arXiv preprint arXiv:2104.02532."},{"issue":"2","key":"1675_CR27","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1109\/MTS.2021.3056293","volume":"40","author":"X Ferrer","year":"2021","unstructured":"Ferrer X, van Nuenen T, Such JM, Cot\u00e9 M, Criado N (2021) Bias and discrimination in AI: a cross-disciplinary perspective. IEEE Technol Soc Mag 40(2):72\u201380","journal-title":"IEEE Technol Soc Mag"},{"issue":"4","key":"1675_CR28","doi-asserted-by":"publisher","first-page":"1473","DOI":"10.1007\/s00146-021-01201-4","volume":"37","author":"F Filgueiras","year":"2022","unstructured":"Filgueiras F (2022) New Pythias of public administration: ambiguity and choice in AI systems as challenges for governance. AI & Soc 37(4):1473\u20131486","journal-title":"AI & Soc"},{"key":"1675_CR29","unstructured":"Font JE, Costa-Jussa MR (2019) Equalizing gender biases in neural machine translation with word embeddings techniques. arXiv preprint https:\/\/arxiv.org\/abs\/1901.03116"},{"key":"1675_CR30","unstructured":"Fountain JE (2004) Building the virtual state: Information technology and institutional change. Brookings Institution Press"},{"issue":"3","key":"1675_CR31","doi-asserted-by":"publisher","first-page":"330","DOI":"10.1145\/230538.230561","volume":"14","author":"B Friedman","year":"1996","unstructured":"Friedman B, Nissenbaum H (1996) Bias in computer systems. ACM Trans Inform Syst (TOIS) 14(3):330\u2013347","journal-title":"ACM Trans Inform Syst (TOIS)"},{"key":"1675_CR32","first-page":"439","volume":"15","author":"M Gutierrez","year":"2021","unstructured":"Gutierrez M (2021) New feminist studies in audiovisual industries | algorithmic gender bias and audiovisual data: a research agenda. Int J Commun 15:439\u2013461","journal-title":"Int J Commun"},{"key":"1675_CR33","doi-asserted-by":"crossref","unstructured":"Hundt A, Agnew W, Zeng V, Kacianka S, Gombolay M (2022, June). Robots Enact Malignant Stereotypes. In 2022 ACM Conference on Fairness, Accountability, and Transparency (pp. 743\u2013756)","DOI":"10.1145\/3531146.3533138"},{"key":"1675_CR34","unstructured":"Igoe, KJ (2021) Algorithmic Bias in Health Care Exacerbates Social Inequities \u2014 How to Prevent It. Harvard T. H Chan School of Public Health. https:\/\/www.hsph.harvard.edu\/ecpe\/how-to-prevent-algorithmic-bias-in-health-care\/. Accessed 17 Jan 2023"},{"key":"1675_CR35","first-page":"1","volume":"3","author":"JY Jang","year":"2019","unstructured":"Jang JY, Lee S, Lee B (2019) Quantification of gender representation bias in commercial films based on image analysis. Proceed ACM on Human-Comput Interact 3:1\u201329","journal-title":"Proceed ACM on Human-Comput Interact"},{"issue":"11","key":"1675_CR36","doi-asserted-by":"publisher","first-page":"e0165037","DOI":"10.1371\/journal.pone.0165037","volume":"11","author":"AH Kerkhoven","year":"2016","unstructured":"Kerkhoven AH, Russo P, Land-Zandstra AM, Saxena A, Rodenburg FJ (2016) Gender stereotypes in science education resources: A visual content analysis. PLoS ONE 11(11):e0165037. https:\/\/doi.org\/10.1371\/journal.pone.0165037","journal-title":"PLoS ONE"},{"key":"1675_CR37","unstructured":"Lee, NT, Resnick P, Barton G (2019) Algorithmic bias detection and mitigation: Best practices and policies to reduce consumer harms. Brookings Institute: Washington, DC, USA. https:\/\/www.brookings.edu\/research\/algorithmic-bias-detection-and-mitigation-best-practices-and-policies-to-reduce-consumer-harms\/. Accessed 17 Jan 2023"},{"key":"1675_CR38","unstructured":"McCarthy, J. (2007). What is artificial intelligence? http:\/\/jmc.stanford.edu\/articles\/whatisai\/whatisai.pdf. Accessed 3 February 2023"},{"key":"1675_CR39","doi-asserted-by":"publisher","unstructured":"Menegatti M, Rubini M (2017) Gender bias and sexism in language. In Oxford Research Encyclopedia of Communication. https:\/\/oxfordre.com\/communication\/view\/https:\/\/doi.org\/10.1093\/acrefore\/9780190228613.001.0001\/acrefore-9780190228613-e-470","DOI":"10.1093\/acrefore\/9780190228613.001.0001\/acrefore-9780190228613-e-470"},{"key":"1675_CR40","unstructured":"MIT Media Lab-a (2018) Gender Shades Project: Frequently Asked Questions. https:\/\/www.media.mit.edu\/projects\/gender-shades\/faq\/#faq-after-a-face-is-detected-what-sort-of-recognition-tasks-can-be-done. Accessed 17 Jan 2023"},{"key":"1675_CR41","unstructured":"MIT Media Lab-b (2018) Gender Shades Project: Why This Matters. https:\/\/www.media.mit.edu\/projects\/gender-shades\/why-this-matters\/. Accessed 17 Jan 2023"},{"key":"1675_CR42","unstructured":"MIT Technology Review (2016). How Vector Space Mathematics Reveals the Hidden Sexism in Language. https:\/\/www.technologyreview.com\/2016\/07\/27\/158634\/how-vector-space-mathematics-reveals-the-hidden-sexism-in-language\/. Accessed 17 Jan 2023"},{"key":"1675_CR43","doi-asserted-by":"crossref","unstructured":"Noble SU (2018) Algorithms of oppression. New York University Press","DOI":"10.2307\/j.ctt1pwt9w5"},{"issue":"3","key":"1675_CR44","doi-asserted-by":"publisher","first-page":"398","DOI":"10.1287\/orsc.3.3.398","volume":"3","author":"WJ Orlikowski","year":"1992","unstructured":"Orlikowski WJ (1992) The duality of technology: Rethinking the concept of technology in organizations. Organ Sci 3(3):398\u2013427","journal-title":"Organ Sci"},{"key":"1675_CR45","unstructured":"Perez CC (2019) Invisible women: Data bias in a world designed for men. Abrams"},{"issue":"10","key":"1675_CR46","doi-asserted-by":"publisher","first-page":"6363","DOI":"10.1007\/s00521-019-04144-6","volume":"32","author":"MO Prates","year":"2020","unstructured":"Prates MO, Avelar PH, Lamb LC (2020) Assessing gender bias in machine translation: a case study with google translate. Neural Comput Appl 32(10):6363\u20136381","journal-title":"Neural Comput Appl"},{"key":"1675_CR47","doi-asserted-by":"crossref","unstructured":"Roselli D, Matthews J, Talagala N (2019) Managing Bias In AI: What Should Businesses Do? Forbes. https:\/\/www.forbes.com\/sites\/cognitiveworld\/2019\/05\/29\/managing-bias-in-ai-what-should-businesses-do\/?sh=10503ad21440. Accessed 17 Jan 2023","DOI":"10.1145\/3308560.3317590"},{"key":"1675_CR48","doi-asserted-by":"publisher","first-page":"81370","DOI":"10.1109\/ACCESS.2021.3084121","volume":"9","author":"T Salazar","year":"2021","unstructured":"Salazar T, Santos MS, Ara\u00fajo H, Abreu PH (2021) FAWOS: fairness-aware oversampling algorithm based on distributions of sensitive attributes. IEEE Access 9:81370\u201381379","journal-title":"IEEE Access"},{"key":"1675_CR49","doi-asserted-by":"publisher","first-page":"845","DOI":"10.1162\/tacl_a_00401","volume":"9","author":"B Savoldi","year":"2021","unstructured":"Savoldi B, Gaido M, Bentivogli L, Negri M, Turchi M (2021) Gender bias in machine translation. Trans Assoc Comput Linguist 9:845\u2013874","journal-title":"Trans Assoc Comput Linguist"},{"key":"1675_CR50","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1177\/2378023120967171","volume":"6","author":"C Schwemmer","year":"2020","unstructured":"Schwemmer C, Knight C, Bello-Pardo ED, Oklobdzija S, Schoonvelde M, Lockhart JW (2020) Diagnosing gender bias in image recognition systems. Socius 6:1\u201317","journal-title":"Socius"},{"key":"1675_CR51","doi-asserted-by":"crossref","unstructured":"Sheng E, Chang KW, Natarajan P, Peng N (2019) The woman worked as a babysitter: On biases in language generation. arXiv preprint arXiv:1909.01326","DOI":"10.18653\/v1\/D19-1339"},{"issue":"11","key":"1675_CR52","first-page":"1281","volume":"71","author":"VK Singh","year":"2020","unstructured":"Singh VK, Chayko M, Inamdar R, Floegel D (2020) Female librarians and male computer programmers? Gender bias in occupational images on digital media platforms. J Am Soc Inf Sci 71(11):1281\u20131294","journal-title":"J Am Soc Inf Sci"},{"key":"1675_CR53","unstructured":"Stafanovi\u010ds A, Bergmanis T, Pinnis M (2020) Mitigating gender bias in machine translation with target gender annotations. arXiv preprint arXiv:2010.06203."},{"key":"1675_CR54","doi-asserted-by":"crossref","unstructured":"Stanovsky G, Smith NA, Zettlemoyer L (2019) Evaluating gender bias in machine translation. arXiv preprint arXiv:1906.00591.","DOI":"10.18653\/v1\/P19-1164"},{"key":"1675_CR55","unstructured":"Stella R (2021) A Dataset for Studying Gender Bias in Translation. Google AI Blog. https:\/\/ai.googleblog.com\/2021\/06\/a-dataset-for-studying-gender-bias-in.html. Accessed 18 Jan 2023"},{"issue":"7781","key":"1675_CR56","doi-asserted-by":"publisher","first-page":"137","DOI":"10.1038\/s41586-019-1657-6","volume":"575","author":"C Tannenbaum","year":"2019","unstructured":"Tannenbaum C, Ellis RP, Eyssel F, Zou J, Schiebinger L (2019) Sex and gender analysis improves science and engineering. Nature 575(7781):137\u2013146","journal-title":"Nature"},{"key":"1675_CR57","unstructured":"TensorFlow (2022) word2vec. https:\/\/www.tensorflow.org\/tutorials\/text\/word2vec. Accessed 17 Jan 2023"},{"issue":"3","key":"1675_CR58","doi-asserted-by":"publisher","first-page":"419","DOI":"10.1007\/s10676-021-09583-1","volume":"23","author":"M Tomalin","year":"2021","unstructured":"Tomalin M, Byrne B, Concannon S, Saunders D, Ullmann S (2021) The practical ethics of bias reduction in machine translation: why domain adaptation is better than data debiasing. Ethics Inf Technol 23(3):419\u2013433","journal-title":"Ethics Inf Technol"},{"key":"1675_CR59","unstructured":"UNESCO (2020, August) Artificial intelligence and gender equality: key findings of UNESCO\u2019s Global Dialogue (Document code: GEN\/2020\/AI\/2 REV). https:\/\/unesdoc.unesco.org\/ark:\/48223\/pf0000374174. Accessed 17 Jan 2023"},{"key":"1675_CR60","doi-asserted-by":"publisher","unstructured":"Vanmassenhove E (2020) On the Integration of Linguistic Features into Statistical and Neural Machine Translation. arXiv preprint arXiv:2003.14324https:\/\/doi.org\/10.48550\/arXiv.2003.14324","DOI":"10.48550\/arXiv.2003.14324"},{"key":"1675_CR61","unstructured":"Vincent J (2020) Service that uses AI to identify gender based on names looks incredibly biased \/ Meghan Smith is a woman, but Dr. Meghan Smith is a man, says Genderify. The Verge. https:\/\/www.theverge.com\/2020\/7\/29\/21346310\/ai-service-gender-verification-identification-genderify. Accessed 17 Jan 2023"},{"key":"1675_CR62","doi-asserted-by":"crossref","unstructured":"Wang T, Zhao J, Yatskar M, Chang KW, Ordonez V (2019) Balanced datasets are not enough: Estimating and mitigating gender bias in deep image representations. In Proceedings of the IEEE\/CVF International Conference on Computer Vision (pp. 5310\u20135319)","DOI":"10.1109\/ICCV.2019.00541"},{"key":"1675_CR63","unstructured":"West DM, Allen JR (2018) How artificial intelligence is transforming the world. Brookings. https:\/\/www.brookings.edu\/research\/how-artificial-intelligence-is-transforming-the-world\/. Accessed 17 Jan 2023"},{"key":"1675_CR64","unstructured":"Zou J (2016) Removing gender bias from algorithms. The Conversation. https:\/\/theconversation.com\/removing-gender-bias-from-algorithms-64721. Accessed 17 Jan 2023"}],"container-title":["AI &amp; SOCIETY"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00146-023-01675-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00146-023-01675-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00146-023-01675-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,13]],"date-time":"2024-08-13T06:16:07Z","timestamp":1723529767000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00146-023-01675-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,5,9]]},"references-count":64,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["1675"],"URL":"https:\/\/doi.org\/10.1007\/s00146-023-01675-4","relation":{},"ISSN":["0951-5666","1435-5655"],"issn-type":[{"value":"0951-5666","type":"print"},{"value":"1435-5655","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,5,9]]},"assertion":[{"value":"3 February 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 April 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 May 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article. Data sharing is not applicable to this article as no datasets were generated or analysed during the current study.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}