{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T16:34:41Z","timestamp":1774542881832,"version":"3.50.1"},"reference-count":64,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2025,8,1]],"date-time":"2025-08-01T00:00:00Z","timestamp":1754006400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,8,1]],"date-time":"2025-08-01T00:00:00Z","timestamp":1754006400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Hum-Cent Intell Syst"],"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>Artificial Intelligence (AI) tools are becoming deeply embedded in everyday life and increasingly influence or automate decision-making processes that could shape not only public opinion but also policies. As their potential impact grows, it is essential to assess the inclusivity of the policy recommendations they could generate and potential biases they may reinforce. This study examines whether AI systems inherently consider gender in policy proposals, both when gender is explicitly mentioned in prompts and when it is not. We conduct four experiments across diverse policy-making contexts to evaluate whether AI-generated recommendations include, overlook, or misrepresent gender considerations. We tested these experiments in two different AI tools, namely ChatGPT (GPT-4) and Microsoft Copilot. To ensure neutrality and reproducibility, we minimize user-specific context and repeat each prompt multiple times. Our findings offer insights into the limitations of current AI tools as policy advisors and contribute to ongoing discussions on algorithmic fairness, implicit gender bias, and the need for gender-aware AI governance. They also raise broader questions about how AI tools understand and represent gender, and how these representations influence the politics of policy-making.  <\/jats:p>","DOI":"10.1007\/s44230-025-00109-2","type":"journal-article","created":{"date-parts":[[2025,8,1]],"date-time":"2025-08-01T10:06:50Z","timestamp":1754042810000},"page":"385-417","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Algorithmic Governance: Gender Bias in AI-Generated Policymaking?"],"prefix":"10.1007","volume":"5","author":[{"given":"Dialekti Athina","family":"Voutyrakou","sequence":"first","affiliation":[]},{"given":"Constantine","family":"Skordoulis","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,1]]},"reference":[{"key":"109_CR1","unstructured":"Dandotiya AS, Gupta SK, Dandotiya N, Sharma MP (2024). AI in everyday life: Transforming society. Navi International Book Publication house."},{"key":"109_CR2","doi-asserted-by":"publisher","unstructured":"Lee RST. Artificial intelligence in daily life. Springer; 2020.\u00a0https:\/\/doi.org\/10.1007\/978-981-15-7695-9.","DOI":"10.1007\/978-981-15-7695-9"},{"key":"109_CR3","doi-asserted-by":"crossref","unstructured":"Bender EM, Gebru T, McMillan-Major A, Shmitchell S (2021). On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (pp. 610\u2013623).","DOI":"10.1145\/3442188.3445922"},{"issue":"3","key":"109_CR4","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1038\/s41746-020-0288-5","volume":"1","author":"D Cirillo","year":"2020","unstructured":"Cirillo D, Catuara-Solarz S, Morey C, Guney E, Subirats L, Mellino S, Gigante A, Valencia A, Rementeria MJ, Chadha AS, Mavridis N. Sex and gender differences and biases in artificial intelligence for biomedicine and healthcare. NPJ Digit Med. 2020;1(3):81. https:\/\/doi.org\/10.1038\/s41746-020-0288-5.","journal-title":"NPJ Digit Med"},{"key":"109_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.4018\/IJIIT.309582","volume":"18","author":"E Kartal","year":"2022","unstructured":"Kartal E. A comprehensive study on bias in artificial intelligence systems: biased or unbiased AI, That\u2019s the Question! Int J Intell Inf Technol. 2022;18:1\u201323.","journal-title":"Int J Intell Inf Technol"},{"key":"109_CR6","doi-asserted-by":"publisher","first-page":"747","DOI":"10.1007\/s00146-022-01474-3","volume":"38","author":"L Marinucci","year":"2023","unstructured":"Marinucci L, Mazzuca C, Gangemi A. Exposing implicit biases and stereotypes in human and artificial intelligence: state of the art and challenges with a focus on gender. AI & Soc. 2023;38:747\u201361.","journal-title":"AI & Soc"},{"key":"109_CR7","first-page":"127","volume":"13","author":"GP Wellner","year":"2020","unstructured":"Wellner GP. When AI is gender-biased. Hum Mente J Philos Stud. 2020;13:127\u201350.","journal-title":"Hum Mente J Philos Stud"},{"key":"109_CR8","doi-asserted-by":"publisher","DOI":"10.3390\/sci6010003","author":"E Ferrara","year":"2023","unstructured":"Ferrara E. Fairness and bias in artificial intelligence: a brief survey of sources, impacts, and mitigation strategies. Sci. 2023. https:\/\/doi.org\/10.3390\/sci6010003.","journal-title":"Sci"},{"key":"109_CR9","doi-asserted-by":"publisher","first-page":"e51837","DOI":"10.2196\/51837","volume":"26","author":"DM Kaplan","year":"2024","unstructured":"Kaplan DM, et al. What\u2019s in a Name? experimental evidence of gender bias in recommendation letters generated by ChatGPT. J Med Internet Res. 2024;26:e51837.","journal-title":"J Med Internet Res"},{"key":"109_CR10","doi-asserted-by":"publisher","first-page":"2045","DOI":"10.1007\/s00146-023-01675-4","volume":"39","author":"S O\u2019Connor","year":"2024","unstructured":"O\u2019Connor S, Liu H. Gender bias perpetuation and mitigation in AI technologies: challenges and opportunities. AI & Soc. 2024;39:2045\u201357. https:\/\/doi.org\/10.1007\/s00146-023-01675-4.","journal-title":"AI & Soc"},{"key":"109_CR11","doi-asserted-by":"crossref","unstructured":"Busker T, Choenni S, Shoae Bargh M. \"Stereotypes in ChatGPT: an empirical study.\" Proceedings of the 16th International Conference on Theory and Practice of Electronic Governance. 2023.","DOI":"10.1145\/3614321.3614325"},{"key":"109_CR12","unstructured":"West SM, Whittaker M, Crawford K (2019) Discriminating Systems: Gender, Race and Power in AI. AI Now Institute."},{"key":"109_CR13","unstructured":"Reuben B. Fairness in Machine Learning: Lessons from Political Philosophy. In Sorelle A. Friedler, Christo Wilson, editors, Conference on Fairness, Accountability and Transparency, FAT 2018, 23\u201324 February 2018, New York, NY, USA. Volume 81 of Proceedings of Machine Learning Research, pages 149\u2013159, PMLR, 2018."},{"key":"109_CR14","doi-asserted-by":"publisher","first-page":"204","DOI":"10.4236\/aasoci.2025.153011","volume":"15","author":"D Voutyrakou","year":"2025","unstructured":"Voutyrakou D, Katsiampoura G, Skordoulis C. Fairness in AI: when are AI tools gender-biased? Adv Appl Sociol. 2025;15:204\u201335. https:\/\/doi.org\/10.4236\/aasoci.2025.153011.","journal-title":"Adv Appl Sociol"},{"key":"109_CR15","first-page":"102775","volume":"76","author":"T Hou","year":"2024","unstructured":"Hou T, Tseng Y, Yuan CW. Is This AI Sexist? The effects of a biased AI\u2019s anthropomorphic appearance and explainability on users\u2019 bias perceptions and trust. Int J Inf Manag. 2024;76:102775.","journal-title":"Int J Inf Manag"},{"key":"109_CR16","doi-asserted-by":"publisher","first-page":"zmad045","DOI":"10.1093\/jcmc\/zmad045","volume":"29","author":"L Sun","year":"2024","unstructured":"Sun L, Wei M, Sun Y, Suh YJ, Shen L, Yang S. Smiling women pitching down: auditing representational and presentational gender biases in image-generative AI. J Comput-Med Commun. 2024;29:zmad045. https:\/\/doi.org\/10.1093\/jcmc\/zmad045.","journal-title":"J Comput-Med Commun"},{"key":"109_CR17","doi-asserted-by":"publisher","first-page":"33","DOI":"10.1007\/s12027-022-00709-6","volume":"23","author":"F L\u00fctz","year":"2022","unstructured":"L\u00fctz F. Gender equality and artificial intelligence in Europe. addressing direct and indirect impacts of algorithms on gender-based discrimination. ERA Forum. 2022;23:33\u201352. https:\/\/doi.org\/10.1007\/s12027-022-00709-6.","journal-title":"ERA Forum"},{"issue":"1","key":"109_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.51357\/jdll.v3i1.218","volume":"3","author":"J Laffier","year":"2023","unstructured":"Laffier J, Rehman A. Deepfakes and harm to women. J Digit Life Learn. 2023;3(1):1\u201321.","journal-title":"J Digit Life Learn"},{"issue":"3","key":"109_CR19","first-page":"52","volume":"6","author":"G Babaei","year":"2024","unstructured":"Babaei G, Banks D, Bosone C, Giudici P, Shan Y. Is ChatGPT more biased than you? Harv Data Sci Rev. 2024;6(3):52.","journal-title":"Harv Data Sci Rev"},{"key":"109_CR20","unstructured":"Urchs S, Thurner V, A\u00dfenmacher M, Heumann C, Thiemichen S \"How Prevalent is Gender Bias in ChatGPT?--Exploring German and English ChatGPT Responses.\" arXiv preprint arXiv:2310.03031 (2023)."},{"key":"109_CR21","doi-asserted-by":"crossref","unstructured":"Fulgu RA, Capraro V. \"Surprising gender biases in GPT.\" arXiv preprint arXiv:2407.06003 (2024).","DOI":"10.2139\/ssrn.4888393"},{"key":"109_CR22","unstructured":"Soundararajan S, Delany SJ, \"Investigating gender bias in large language models through text generation,\" in Proceedings of the 7th International Conference on Natural Language and Speech Processing (ICNLSP 2024), 2024, pp. 410\u2013424."},{"key":"109_CR23","unstructured":"Lorentzen B (2022). Social Biases in Language Models\u202f: Gender Stereotypes in GPT-3 Generated Stories (Dissertation)."},{"key":"109_CR24","doi-asserted-by":"crossref","unstructured":"Kotek H, Dockum R, Sun D, \"Gender bias and stereotypes in large language models,\" in Proceedings of the ACM Collective Intelligence Conference, 2023, pp. 12\u201324.","DOI":"10.1145\/3582269.3615599"},{"key":"109_CR25","doi-asserted-by":"crossref","unstructured":"Ghosh S, Caliskan A (2023). Chatgpt perpetuates gender bias in machine translation and ignores non-gendered pronouns: Findings across bengali and five other low-resource languages. In Proceedings of the 2023 AAAI\/ACM Conference on AI, Ethics, and Society (pp. 901\u2013912).","DOI":"10.1145\/3600211.3604672"},{"key":"109_CR26","unstructured":"Zhou KZ, Sanfilippo MR (2023). Public perceptions of gender bias in large language models: Cases of chatgpt and ernie. arXiv preprint arXiv:2309.09120."},{"issue":"4","key":"109_CR27","doi-asserted-by":"publisher","first-page":"1032","DOI":"10.55214\/25768484.v9i4.6169","volume":"9","author":"DA Voutyrakou","year":"2025","unstructured":"Voutyrakou DA, Katsiampoura G, Skordoulis C. Are AI tools biased storytellers? Examining gender-bias in AI-generated narratives. Edelweiss Appl Sci Technol. 2025;9(4):1032\u201349. https:\/\/doi.org\/10.55214\/25768484.v9i4.6169.","journal-title":"Edelweiss Appl Sci Technol"},{"key":"109_CR28","doi-asserted-by":"crossref","unstructured":"Bano M, Gunatilake H, Hoda R (2025). What Does a Software Engineer Look Like? Exploring Societal Stereotypes in LLMs. arXiv preprint arXiv:2501.03569.","DOI":"10.1109\/ICSE-SEIS66351.2025.00023"},{"issue":"5","key":"109_CR29","doi-asserted-by":"publisher","first-page":"250","DOI":"10.3390\/socsci13050250","volume":"13","author":"T Sandoval-Martin","year":"2024","unstructured":"Sandoval-Martin T, Mart\u00ednez-Sanzo E. Perpetuation of gender bias in visual representation of professions in the generative ai tools dall\u00b7 e and bing image creator. Soc Sci. 2024;13(5):250.","journal-title":"Soc Sci"},{"key":"109_CR30","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1080\/09718524.2022.2128254","volume":"26","author":"A Manasi","year":"2022","unstructured":"Manasi A, Panchanadeswaran S, Sours E, Lee SJ. Mirroring the bias: gender and artificial intelligence. Gend Technol Dev. 2022;26:295\u2013305.","journal-title":"Gend Technol Dev"},{"key":"109_CR31","doi-asserted-by":"publisher","first-page":"1264","DOI":"10.1108\/OIR-08-2021-0452","volume":"47","author":"P Hall","year":"2023","unstructured":"Hall P, Ellis D. A systematic review of socio-technical gender bias in AI algorithms. Online Inf Rev. 2023;47:1264\u201379.","journal-title":"Online Inf Rev"},{"key":"109_CR32","doi-asserted-by":"publisher","unstructured":"Berge J. J. C. Nunnally and I. H. Bernstein. Psychometric Theory. New York: McGraw-Hill, 1994, xxiv+752 pages,\u00a0\u00a351.95. Psychometrika 2025;60:313-315. https:\/\/doi.org\/10.1007\/BF02301419.","DOI":"10.1007\/BF02301419"},{"key":"109_CR33","volume-title":"Multivariate data analysis","author":"JF Hair","year":"2010","unstructured":"Hair JF, Black WC, Babin BJ, Anderson RE. Multivariate data analysis. 7th ed. New York: Pearson; 2010.","edition":"7"},{"key":"109_CR34","doi-asserted-by":"publisher","first-page":"53","DOI":"10.5116\/ijme.4dfb.8dfd","volume":"2","author":"M Tavakol","year":"2011","unstructured":"Tavakol M, Dennick R. Making sense of Cronbach\u2019s alpha. Int J Med Educ. 2011;2:53\u20135. https:\/\/doi.org\/10.5116\/ijme.4dfb.8dfd.","journal-title":"Int J Med Educ"},{"key":"109_CR35","first-page":"370","volume":"73","author":"E Montano","year":"2018","unstructured":"Montano E. The bring your own tampon policy: Why menstrual hygiene products should be provided for free in restrooms. U Miami L Rev. 2018;73:370.","journal-title":"U Miami L Rev"},{"key":"109_CR36","first-page":"1","volume":"53","author":"ME Johnson","year":"2019","unstructured":"Johnson ME. Menstrual justice. UC Davis L Rev. 2019;53:1.","journal-title":"UC Davis L Rev"},{"key":"109_CR37","unstructured":"Perez CC. Invisible women: Data bias in a world designed for men. Abrams; 2019."},{"issue":"12","key":"109_CR38","doi-asserted-by":"publisher","first-page":"1054","DOI":"10.1038\/nclimate2741","volume":"5","author":"B Kingma","year":"2015","unstructured":"Kingma B, van Marken Lichtenbelt WD. Energy consumption in buildings and female thermal demand. Nat Clim Chang. 2015;5(12):1054\u20136. https:\/\/doi.org\/10.1038\/nclimate2741.","journal-title":"Nat Clim Chang"},{"key":"109_CR39","volume-title":"The second shift: working parents and the revolution at home","author":"AR Hochschild","year":"1989","unstructured":"Hochschild AR. The second shift: working parents and the revolution at home. New York: Viking; 1989."},{"key":"109_CR40","volume-title":"Gender in world perspective","author":"RW Connell","year":"2009","unstructured":"Connell RW. Gender in world perspective. Cambridge: Polity Press; 2009."},{"key":"109_CR41","volume-title":"Feminist city","author":"L Kern","year":"2021","unstructured":"Kern L. Feminist city. London: Verso Books; 2021."},{"issue":"9","key":"109_CR42","doi-asserted-by":"publisher","first-page":"1217","DOI":"10.1080\/02723638.2020.1732847","volume":"41","author":"L Kern","year":"2020","unstructured":"Kern L. Feminist spatial praxis: toward an ethics of care in urban design. Urban Geogr. 2020;41(9):1217\u201337. https:\/\/doi.org\/10.1080\/02723638.2020.1732847.","journal-title":"Urban Geogr"},{"key":"109_CR43","doi-asserted-by":"publisher","unstructured":"Bobel C, editor. The managed body: developing girls and menstrual health in the global south. Cham: Springer;\u00a02019.\u00a0https:\/\/doi.org\/10.1007\/978-3-319-89414-0_1.","DOI":"10.1007\/978-3-319-89414-0_1"},{"issue":"1","key":"109_CR44","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1007\/s11199-011-0052-z","volume":"68","author":"I Johnston-Robledo","year":"2020","unstructured":"Johnston-Robledo I, Chrisler JC. The menstrual mark: menstruation as social stigma. Sex Roles. 2020;68(1):9\u201318. https:\/\/doi.org\/10.1007\/s11199-011-0052-z.","journal-title":"Sex Roles"},{"key":"109_CR45","doi-asserted-by":"publisher","DOI":"10.2979\/3270.0","volume-title":"Volatile bodies","author":"EA Grosz","year":"1994","unstructured":"Grosz EA. Volatile bodies. Bloomington: Indiana University Press; 1994."},{"issue":"3","key":"109_CR46","doi-asserted-by":"publisher","first-page":"301","DOI":"10.1080\/09663699725369","volume":"4","author":"H Koskela","year":"1997","unstructured":"Koskela H. \u201cBold Walk and Breakings\u201d: women\u2019s spatial confidence versus fear of violence. Gend Place Cult. 1997;4(3):301\u201320. https:\/\/doi.org\/10.1080\/09663699725369.","journal-title":"Gend Place Cult"},{"key":"109_CR47","volume-title":"Space, Place, and Gender","author":"D Massey","year":"1994","unstructured":"Massey D. Space, Place, and Gender. University of Minnesota Press; 1994."},{"key":"109_CR48","doi-asserted-by":"crossref","unstructured":"Crenshaw K (1991). Mapping the margins: Intersectionality, identity politics, and violence against women of color. Stanford Law Review, 43(6), 1241\u20131299","DOI":"10.2307\/1229039"},{"key":"109_CR49","volume-title":"Artificial unintelligence: how computers misunderstand the world","author":"M Broussard","year":"2019","unstructured":"Broussard M. Artificial unintelligence: how computers misunderstand the world. Cambridge and London: MIT Press; 2019."},{"key":"109_CR50","doi-asserted-by":"publisher","first-page":"1146260","DOI":"10.3389\/fpsyg.2023.1146260","volume":"14","author":"X Xie","year":"2023","unstructured":"Xie X, Wang R, Gou Z, Chen S. A comparative study of gender differences in healthy office building strategies. Front Psychol. 2023;14:1146260.","journal-title":"Front Psychol"},{"key":"109_CR51","volume-title":"The second shift: Working families and the revolution at home","author":"A Hochschild","year":"1989","unstructured":"Hochschild A, Machung A. The second shift: Working families and the revolution at home. London: Penguin Books; 1989."},{"issue":"6","key":"109_CR52","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3457607","volume":"54","author":"N Mehrabi","year":"2021","unstructured":"Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A. A survey on bias and fairness in machine learning. ACM Comput Surv (CSUR). 2021;54(6):1\u201335. https:\/\/doi.org\/10.1145\/3457607.","journal-title":"ACM Comput Surv (CSUR)"},{"key":"109_CR53","volume-title":"Moral boundaries: a political argument for an ethic of care","author":"J Tronto","year":"1993","unstructured":"Tronto J. Moral boundaries: a political argument for an ethic of care. New York & London: Routledge; 1993."},{"key":"109_CR54","doi-asserted-by":"publisher","DOI":"10.1093\/0198297556.001.0001","volume-title":"Inclusion and democracy","author":"IM Young","year":"2002","unstructured":"Young IM. Inclusion and democracy. New York: Oxford University Press; 2002."},{"key":"109_CR55","volume-title":"Development as freedom","author":"A Sen","year":"1999","unstructured":"Sen A. Development as freedom. Oxford: Oxford University Press; 1999."},{"key":"109_CR56","doi-asserted-by":"publisher","DOI":"10.4159\/harvard.9780674061200","volume-title":"Creating capabilities: the human development approach","author":"MC Nussbaum","year":"2011","unstructured":"Nussbaum MC. Creating capabilities: the human development approach. USA: Harvard University Press; 2011."},{"key":"109_CR57","doi-asserted-by":"publisher","DOI":"10.1093\/0195180992.001.0001","volume-title":"The ethics of care: Personal, political, and global","author":"V Held","year":"2005","unstructured":"Held V. The ethics of care: Personal, political, and global. Oxford University Press; 2005."},{"key":"109_CR58","doi-asserted-by":"crossref","unstructured":"D'ignazio C, Klein LF. Data feminism. Cambridge, MA: MIT press; 2020.","DOI":"10.7551\/mitpress\/11805.001.0001"},{"key":"109_CR59","doi-asserted-by":"publisher","DOI":"10.7551\/mitpress\/12255.001.0001","volume-title":"Design justice: community-led practices to build the worlds we need","author":"S Costanza-Chock","year":"2020","unstructured":"Costanza-Chock S. Design justice: community-led practices to build the worlds we need. Cambridge: The MIT Press; 2020."},{"key":"109_CR60","doi-asserted-by":"crossref","unstructured":"Raji I, Smart A, White R, Mitchell M, Gebru T, Hutchinson B, Smith-Loud J, Theron D, Barnes P. (2020). Closing the AI accountability gap: defining an end-to-end framework for internal algorithmic auditing. 33\u201344","DOI":"10.1145\/3351095.3372873"},{"key":"109_CR61","unstructured":"Buolamwini J, Gebru T (2018) Gender shades: Intersectional accuracy disparities in commercial gender classification. Proceedings of the 1st Conference on Fairness, Accountability and Transparency, PMLR 81:77\u201391. https:\/\/proceedings.mlr.press\/v81\/buolamwini18a.html"},{"key":"109_CR62","doi-asserted-by":"crossref","unstructured":"Amershi S, Weld D, Vorvoreanu M, Fourney A, Nushi B, Collisson P, Suh J, Iqbal S, Bennett PN, Inkpen K, Teevan J, Kikin-Gil R, Horvitz E. 2019. Guidelines for Human-AI Interaction. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (CHI '19). Association for Computing Machinery, New York, NY, USA, Paper 3, 1\u201313","DOI":"10.1145\/3290605.3300233"},{"key":"109_CR63","doi-asserted-by":"publisher","unstructured":"Doshi-Velez F, Kim B (2017) Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608. https:\/\/doi.org\/10.48550\/arXiv.1702.08608","DOI":"10.48550\/arXiv.1702.08608"},{"issue":"1","key":"109_CR64","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1038\/s42256-019-0138-9","volume":"2","author":"SM Lundberg","year":"2020","unstructured":"Lundberg SM, Erion G, Chen H, DeGrave A, Prutkin JM, Nair B, Katz R, Himmelfarb J, Bansal N, Lee SI. From local explanations to global understanding with explainable AI for trees. Nat Mach Intell. 2020;2(1):56\u201367. https:\/\/doi.org\/10.1038\/s42256-019-0138-9.","journal-title":"Nat Mach Intell"}],"container-title":["Human-Centric Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44230-025-00109-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44230-025-00109-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44230-025-00109-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,19]],"date-time":"2025-09-19T10:23:13Z","timestamp":1758277393000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44230-025-00109-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,1]]},"references-count":64,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2025,9]]}},"alternative-id":["109"],"URL":"https:\/\/doi.org\/10.1007\/s44230-025-00109-2","relation":{},"ISSN":["2667-1336"],"issn-type":[{"value":"2667-1336","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,8,1]]},"assertion":[{"value":"9 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 July 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 July 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 August 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}