{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T13:43:17Z","timestamp":1774964597799,"version":"3.50.1"},"reference-count":63,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2023,4,15]],"date-time":"2023-04-15T00:00:00Z","timestamp":1681516800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,4,15]],"date-time":"2023-04-15T00:00:00Z","timestamp":1681516800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Data Sci. Eng."],"published-print":{"date-parts":[[2023,6]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>The ever-increasing number of systems based on semantic text analysis is making natural language understanding a fundamental task: embedding-based language models are used for a variety of applications, such as resume parsing or improving web search results. At the same time, despite their popularity and widespread use, concern is rapidly growing due to their display of social bias and lack of transparency. In particular, they exhibit a large amount of gender bias, favouring the consolidation of social stereotypes. Recently, sentence embeddings have been introduced as a novel and powerful technique to represent entire sentences as vectors. We propose a new metric to estimate gender bias in sentence embeddings, named <jats:italic>bias score<\/jats:italic>. Our solution leverages semantic importance of words and previous research on bias in word embeddings, and it is able to discern between neutral and biased gender information at sentence level. Experiments on a real-world dataset demonstrate that our novel metric can identify gender stereotyped sentences. Furthermore, we employ <jats:italic>bias score<\/jats:italic> to detect and then remove or compensate for the more stereotyped entries in text corpora used to train sentence encoders, improving their degree of fairness. Finally, we prove that models retrained on fairer corpora are less prone to make stereotypical associations compared to their original counterpart, while preserving accuracy in natural language understanding tasks. Additionally, we compare our experiments with traditional methods for reducing bias in embedding-based language models.<\/jats:p>","DOI":"10.1007\/s41019-023-00211-0","type":"journal-article","created":{"date-parts":[[2023,4,15]],"date-time":"2023-04-15T17:02:21Z","timestamp":1681578141000},"page":"177-195","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Improving Gender-Related Fairness in Sentence Encoders: A Semantics-Based Approach"],"prefix":"10.1007","volume":"8","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1403-7766","authenticated-orcid":false,"given":"Tommaso","family":"Dolci","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0631-2120","authenticated-orcid":false,"given":"Fabio","family":"Azzalini","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7172-0203","authenticated-orcid":false,"given":"Mara","family":"Tanelli","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,15]]},"reference":[{"key":"211_CR1","doi-asserted-by":"crossref","unstructured":"Agirre E, Banea C, Cardie C, et\u00a0al (2014) Semeval-2014 task 10: Multilingual semantic textual similarity. In: Proceedings of the 8th international workshop on semantic evaluation (SemEval 2014), pp 81\u201391","DOI":"10.3115\/v1\/S14-2010"},{"key":"211_CR2","doi-asserted-by":"crossref","unstructured":"Basta C, Costa-juss\u00e0 MR, Casas N (2019) Evaluating the underlying gender bias in contextualized word embeddings. In: Proceedings of the first workshop on gender bias in natural language processing (GeBNLP), pp 33\u201339","DOI":"10.18653\/v1\/W19-3805"},{"key":"211_CR3","doi-asserted-by":"publisher","first-page":"587","DOI":"10.1162\/tacl_a_00041","volume":"6","author":"EM Bender","year":"2018","unstructured":"Bender EM, Friedman B (2018) Data statements for natural language processing: toward mitigating system bias and enabling better science. Transact Associat Comput Linguist (TACL) 6:587\u2013604","journal-title":"Transact Associat Comput Linguist (TACL)"},{"key":"211_CR4","doi-asserted-by":"crossref","unstructured":"Bender EM, Gebru T, McMillan-Major A, et\u00a0al (2021) On the dangers of stochastic parrots: Can language models be too big? In: Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (FAccT), pp 610\u2013623","DOI":"10.1145\/3442188.3445922"},{"key":"211_CR5","unstructured":"Bhatia V, Rawat P, Kumar A, et\u00a0al (2019) End-to-end resume parsing and finding candidates for a job description using bert. arXiv preprint arXiv:1910.03089"},{"key":"211_CR6","doi-asserted-by":"crossref","unstructured":"Blodgett SL, Barocas S, Daum\u00e9\u00a0III H, et\u00a0al (2020) Language (technology) is power: A critical survey of \u201cbias\u201d in nlp. In: Proceedings of the 58th annual meeting of the association for computational linguistics (ACL), pp 5454\u20135476","DOI":"10.18653\/v1\/2020.acl-main.485"},{"key":"211_CR7","unstructured":"Bolukbasi T, Chang KW, Zou JY, et\u00a0al (2016) Man is to computer programmer as woman is to homemaker? debiasing word embeddings. Adv Neural Inform Process Syst. 29"},{"key":"211_CR8","doi-asserted-by":"crossref","unstructured":"Bordia S, Bowman SR (2019) Identifying and reducing gender bias in word-level language models. In: Proceedings of the 2019 Conference of the north american chapter of the association for computational linguistics: Human Language Technologies - Student Research Workshop (NAACL-HLT SRW), pp 7\u201315","DOI":"10.18653\/v1\/N19-3002"},{"key":"211_CR9","doi-asserted-by":"crossref","unstructured":"Bowman SR, Angeli G, Potts C, et\u00a0al (2015) A large annotated corpus for learning natural language inference. In: Proceedings of the 2015 conference on empirical methods in natural language processing (EMNLP)","DOI":"10.18653\/v1\/D15-1075"},{"key":"211_CR10","unstructured":"Brunet ME, Alkalay-Houlihan C, Anderson A, et\u00a0al (2019) Understanding the origins of bias in word embeddings. In: International conference on machine learning (ICML), pp 803\u2013811"},{"issue":"6334","key":"211_CR11","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1126\/science.aal4230","volume":"356","author":"A Caliskan","year":"2017","unstructured":"Caliskan A, Bryson JJ, Narayanan A (2017) Semantics derived automatically from language corpora contain human-like biases. Science 356(6334):183\u2013186","journal-title":"Science"},{"key":"211_CR12","doi-asserted-by":"crossref","unstructured":"Cer D, Diab M, Agirre E, et\u00a0al (2017) Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation. arXiv preprint arXiv:1708.00055","DOI":"10.18653\/v1\/S17-2001"},{"key":"211_CR13","doi-asserted-by":"crossref","unstructured":"Cer D, Yang Y, Kong Sy, et\u00a0al (2018) Universal sentence encoder. arXiv preprint arXiv:1803.11175","DOI":"10.18653\/v1\/D18-2029"},{"key":"211_CR14","unstructured":"Chang KW, Prabhakaran V, Ordonez V (2019) Bias and fairness in natural language processing. In: Proceedings of the 2019 conference on empirical methods in natural language processing and 9th international joint conference on natural language processing (EMNLP-IJCNLP): Tutorial Abstracts"},{"key":"211_CR15","unstructured":"Cheng P, Hao W, Yuan S, et\u00a0al (2021) Fairfil: Contrastive neural debiasing method for pretrained text encoders. In: International conference on learning representations (ICLR)"},{"key":"211_CR16","unstructured":"Conneau A, Kiela D (2018) Senteval: An evaluation toolkit for universal sentence representations. In: Proceedings of the eleventh international conference on language resources and evaluation (LREC)"},{"key":"211_CR17","doi-asserted-by":"crossref","unstructured":"Conneau A, Kiela D, Schwenk H, et\u00a0al (2017) Supervised learning of universal sentence representations from natural language inference data. In: Proceedings of the 2017 conference on empirical methods in natural language processing (EMNLP), pp 670\u2013680","DOI":"10.18653\/v1\/D17-1070"},{"key":"211_CR18","unstructured":"Devlin J, Chang MW, Lee K, et\u00a0al (2019) Bert: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 conference of the north american chapter of the association for computational linguistics: Human Language Technologies (NAACL-HLT), pp 4171\u20134186"},{"key":"211_CR19","doi-asserted-by":"crossref","unstructured":"Dolan W, Quirk C, Brockett C, et\u00a0al (2004) Unsupervised construction of large paraphrase corpora: Exploiting massively parallel news sources. In: Proceedings of the 20th international conference on computational linguistics (COLING)","DOI":"10.3115\/1220355.1220406"},{"key":"211_CR20","doi-asserted-by":"crossref","unstructured":"Dolci T (2022) Fine-tuning language models to mitigate gender bias in sentence encoders. In: 2022 IEEE eighth international conference on big data computing service and applications (BigDataService), IEEE, pp 175\u2013176","DOI":"10.1109\/BigDataService55688.2022.00036"},{"key":"211_CR21","doi-asserted-by":"crossref","unstructured":"Du Y, Fang Q, Nguyen D (2021) Assessing the reliability of word embedding gender bias measures. In: Proceedings of the 2021 conference on empirical methods in natural language processing (EMNLP), pp 10,012\u201310,034","DOI":"10.18653\/v1\/2021.emnlp-main.785"},{"issue":"7","key":"211_CR22","doi-asserted-by":"publisher","first-page":"3184","DOI":"10.3390\/app11073184","volume":"11","author":"I Garrido-Mu\u00f1oz","year":"2021","unstructured":"Garrido-Mu\u00f1oz I, Montejo-R\u00e1ez A, Mart\u00ednez-Santiago F et al (2021) A survey on bias in deep nlp. Appl Sci 11(7):3184","journal-title":"Appl Sci"},{"key":"211_CR23","doi-asserted-by":"crossref","unstructured":"Gira M, Zhang R, Lee K (2022) Debiasing pre-trained language models via efficient fine-tuning. In: Proceedings of the second workshop on language technology for equality, diversity and inclusion (LT-EDI), pp 59\u201369","DOI":"10.18653\/v1\/2022.ltedi-1.8"},{"key":"211_CR24","unstructured":"Gonen H, Goldberg Y (2019) Lipstick on a pig: Debiasing methods cover up systematic gender biases in word embeddings but do not remove them. In: Proceedings of the 2019 conference of the north american chapter of the association for computational linguistics: human language technologies (NAACL-HLT), pp 609\u2013614"},{"key":"211_CR25","first-page":"658","volume":"2022","author":"U Gupta","year":"2022","unstructured":"Gupta U, Dhamala J, Kumar V et al (2022) Mitigating gender bias in distilled language models via counterfactual role reversal. Find Associat Comput Linguist: ACL 2022:658\u2013678","journal-title":"Find Associat Comput Linguist: ACL"},{"key":"211_CR26","doi-asserted-by":"crossref","unstructured":"Hu M, Liu B (2004) Mining and summarizing customer reviews. In: Proceedings of the tenth ACM SIGKDD international conference on knowledge discovery and data mining (KDD), pp 168\u2013177","DOI":"10.1145\/1014052.1014073"},{"key":"211_CR27","doi-asserted-by":"crossref","unstructured":"Kaneko M, Bollegala D (2019) Gender-preserving debiasing for pre-trained word embeddings. In: proceedings of the 57th annual meeting of the association for computational linguistics (ACL), pp 1641\u20131650","DOI":"10.18653\/v1\/P19-1160"},{"key":"211_CR28","doi-asserted-by":"crossref","unstructured":"Kaneko M, Bollegala D (2021) Debiasing pre-trained contextualised embeddings. In: Proceedings of the 16th conference of the european chapter of the association for computational linguistics (EACL), pp 1256\u20131266","DOI":"10.18653\/v1\/2021.eacl-main.107"},{"key":"211_CR29","doi-asserted-by":"crossref","unstructured":"Lan Z, Chen M, Goodman S, et\u00a0al (2020) Albert: A lite bert for self-supervised learning of language representations. In: 8th international conference on learning representations (ICLR)","DOI":"10.1109\/SLT48900.2021.9383575"},{"key":"211_CR30","doi-asserted-by":"crossref","unstructured":"Liang PP, Li IM, Zheng E, et\u00a0al (2020) Towards debiasing sentence representations. In: Proceedings of the 58th annual meeting of the association for computational linguistics (ACL), pp 5502\u20135515","DOI":"10.18653\/v1\/2020.acl-main.488"},{"key":"211_CR31","unstructured":"Liu Y, Ott M, Goyal N, et\u00a0al (2019) Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692"},{"key":"211_CR32","doi-asserted-by":"crossref","unstructured":"Lu K, Mardziel P, Wu F, et\u00a0al (2020) Gender bias in neural natural language processing. In: logic, language, and security. Springer, p 189\u2013202","DOI":"10.1007\/978-3-030-62077-6_14"},{"key":"211_CR33","doi-asserted-by":"crossref","unstructured":"MacAvaney S, Yates A, Cohan A, et\u00a0al (2019) Cedr: Contextualized embeddings for document ranking. In: Proceedings of the 42nd international ACM SIGIR conference on research and development in information retrieval, pp 1101\u20131104","DOI":"10.1145\/3331184.3331317"},{"key":"211_CR34","doi-asserted-by":"crossref","unstructured":"Manzini T, Chong LY, Black AW, et\u00a0al (2019) Black is to criminal as caucasian is to police: Detecting and removing multiclass bias in word embeddings. In: Proceedings of the 2019 conference of the north american chapter of the association for computational linguistics: human language technologies, pp 615\u2013621","DOI":"10.18653\/v1\/N19-1062"},{"key":"211_CR35","unstructured":"Marelli M, Menini S, Baroni M, et\u00a0al (2014) A sick cure for the evaluation of compositional distributional semantic models. In: Proceedings of the ninth international conference on language resources and evaluation (LREC), pp 216\u2013223"},{"key":"211_CR36","doi-asserted-by":"crossref","unstructured":"May C, Wang A, Bordia S, et\u00a0al (2019) On measuring social biases in sentence encoders. In: Proceedings of the 2019 conference of the north american chapter of the association for computational linguistics: human language technologies (NAACL-HLT), pp 622\u2013628","DOI":"10.18653\/v1\/N19-1063"},{"key":"211_CR37","doi-asserted-by":"crossref","unstructured":"Meade N, Poole-Dayan E, Reddy S (2022) An empirical survey of the effectiveness of debiasing techniques for pre-trained language models. In: Proceedings of the 60th annual meeting of the association for computational linguistics (ACL), pp 1878\u20131898","DOI":"10.18653\/v1\/2022.acl-long.132"},{"key":"211_CR38","unstructured":"Mikolov T, Chen K, Corrado G, et\u00a0al (2013a) Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781"},{"key":"211_CR39","unstructured":"Mikolov T, Sutskever I, Chen K, et\u00a0al (2013b) Distributed representations of words and phrases and their compositionality. Adv Neural Inform Process Syst 26"},{"key":"211_CR40","unstructured":"Mikolov T, Yih WT, Zweig G (2013c) Linguistic regularities in continuous space word representations. In: Proceedings of the 2013 conference of the north american chapter of the association for computational linguistics: human language technologies (NAACL-HLT)"},{"key":"211_CR41","doi-asserted-by":"crossref","unstructured":"Pang B, Lee L (2004) A sentimental education: Sentiment analysis using subjectivity summarization based on minimum cuts. In: Proceedings of the 42nd annual meeting of the association for computational linguistics (ACL), pp 271\u2013278","DOI":"10.3115\/1218955.1218990"},{"key":"211_CR42","doi-asserted-by":"crossref","unstructured":"Pang B, Lee L (2005) Seeing stars: Exploiting class relationships for sentiment categorization with respect to rating scales. In: Proceedings of the 43rd annual meeting of the association for computational linguistics (ACL), pp 115\u2013124","DOI":"10.3115\/1219840.1219855"},{"key":"211_CR43","doi-asserted-by":"crossref","unstructured":"Park JH, Shin J, Fung P (2018) Reducing gender bias in abusive language detection. In: proceedings of the 2018 conference on empirical methods in natural language processing (EMNLP)","DOI":"10.18653\/v1\/D18-1302"},{"key":"211_CR44","doi-asserted-by":"crossref","unstructured":"Park S, Shim HS, Chatterjee M, et\u00a0al (2014) Computational analysis of persuasiveness in social multimedia: A novel dataset and multimodal prediction approach. In: Proceedings of the 16th international conference on multimodal interaction, pp 50\u201357","DOI":"10.1145\/2663204.2663260"},{"key":"211_CR45","unstructured":"Parraga O, More MD, Oliveira CM, et\u00a0al (2022) Debiasing methods for fairer neural models in vision and language research: A survey. arXiv preprint arXiv:2211.05617"},{"key":"211_CR46","doi-asserted-by":"crossref","unstructured":"Pennington J, Socher R, Manning CD (2014) Glove: Global vectors for word representation. In: Proceedings of the 2014 Conference on empirical methods in natural language processing (EMNLP), pp 1532\u20131543","DOI":"10.3115\/v1\/D14-1162"},{"key":"211_CR47","doi-asserted-by":"crossref","unstructured":"Poria S, Hazarika D, Majumder N, et\u00a0al (2019) Meld: A multimodal multi-party dataset for emotion recognition in conversations. In: Proceedings of the 57th annual meeting of the association for computational linguistics, pp 527\u2013536","DOI":"10.18653\/v1\/P19-1050"},{"key":"211_CR48","doi-asserted-by":"crossref","unstructured":"Qian Y, Muaz U, Zhang B, et\u00a0al (2019) Reducing gender bias in word-level language models with a gender-equalizing loss function. In: Proceedings of the 57th Annual meeting of the association for computational linguistics: student research workshop (ACL SRW), pp 223\u2013228","DOI":"10.18653\/v1\/P19-2031"},{"key":"211_CR49","doi-asserted-by":"crossref","unstructured":"Reimers N, Gurevych I (2019) Sentence-bert: Sentence embeddings using siamese bert-networks. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pp 3982\u20133992","DOI":"10.18653\/v1\/D19-1410"},{"key":"211_CR50","doi-asserted-by":"crossref","unstructured":"Reimers N, Gurevych I (2020) Making monolingual sentence embeddings multilingual using knowledge distillation. In: Proceedings of the 2020 Conference on empirical methods in natural language processing (EMNLP). association for computational linguistics, arXiv: 2004.09813","DOI":"10.18653\/v1\/2020.emnlp-main.365"},{"key":"211_CR51","doi-asserted-by":"crossref","unstructured":"Rudinger R, May C, Van\u00a0Durme B (2017) Social bias in elicited natural language inferences. In: Proceedings of the first workshop on ethics in natural language processing (EthNLP), pp 74\u201379","DOI":"10.18653\/v1\/W17-1609"},{"key":"211_CR52","unstructured":"Sanh V, Debut L, Chaumond J, et\u00a0al (2019) Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108"},{"key":"211_CR53","unstructured":"Socher R, Perelygin A, Wu J, et\u00a0al (2013) Recursive deep models for semantic compositionality over a sentiment treebank. In: Proceedings of the 2013 conference on empirical methods in natural language processing (EMNLP), pp 1631\u20131642"},{"key":"211_CR54","doi-asserted-by":"crossref","unstructured":"Stoyanovich J, Howe B, Jagadish H (2020) Responsible data management. Proceedings of the VLDB Endowment 13(12)","DOI":"10.14778\/3415478.3415570"},{"key":"211_CR55","doi-asserted-by":"crossref","unstructured":"Voorhees EM, Tice DM (2000) Building a question answering test collection. In: Proceedings of the 23rd annual international ACM SIGIR conference on Research and development in information retrieval, pp 200\u2013207","DOI":"10.1145\/345508.345577"},{"key":"211_CR56","doi-asserted-by":"crossref","unstructured":"Wang T, Lin XV, Rajani NF, et\u00a0al (2020) Double-hard debias: Tailoring word embeddings for gender bias mitigation. In: Proceedings of the 58th annual meeting of the association for computational linguistics (ACL), pp 5443\u20135453","DOI":"10.18653\/v1\/2020.acl-main.484"},{"issue":"2","key":"211_CR57","doi-asserted-by":"publisher","first-page":"165","DOI":"10.1007\/s10579-005-7880-9","volume":"39","author":"J Wiebe","year":"2005","unstructured":"Wiebe J, Wilson T, Cardie C (2005) Annotating expressions of opinions and emotions in language. Language Resour Evaluat 39(2):165\u2013210","journal-title":"Language Resour Evaluat"},{"key":"211_CR58","doi-asserted-by":"crossref","unstructured":"Williams A, Nangia N, Bowman S (2018) A broad-coverage challenge corpus for sentence understanding through inference. In: Proceedings of the 2018 conference of the north american chapter of the association for computational linguistics: human language technologies, Vol 1 (Long Papers), pp 1112\u20131122","DOI":"10.18653\/v1\/N18-1101"},{"key":"211_CR59","unstructured":"Xu G, Hu Q (2022) Can model compression improve nlp fairness. arXiv preprint arXiv:2201.08542"},{"key":"211_CR60","doi-asserted-by":"crossref","unstructured":"Yang Z, Feng J (2020) A causal inference method for reducing gender bias in word embedding relations. In: Proceedings of the AAAI conference on artificial intelligence, pp 9434\u20139441","DOI":"10.1609\/aaai.v34i05.6486"},{"key":"211_CR61","unstructured":"Ye W, Xu F, Huang Y, et\u00a0al (2021) Adversarial examples generation for reducing implicit gender bias in pre-trained models. arXiv preprint arXiv:2110.01094"},{"key":"211_CR62","doi-asserted-by":"crossref","unstructured":"Zhao J, Wang T, Yatskar M, et\u00a0al (2018a) Gender bias in coreference resolution: Evaluation and debiasing methods. In: Proceedings of the 2018 conference of the north american chapter of the association for computational linguistics: human language technologies (NAACL-HLT), pp 15\u201320","DOI":"10.18653\/v1\/N18-2003"},{"key":"211_CR63","doi-asserted-by":"crossref","unstructured":"Zhao J, Zhou Y, Li Z, et\u00a0al (2018b) Learning gender-neutral word embeddings. In: proceedings of the 2018 conference on empirical methods in natural language processing (EMNLP), pp 4847\u20134853","DOI":"10.18653\/v1\/D18-1521"}],"container-title":["Data Science and Engineering"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s41019-023-00211-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s41019-023-00211-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s41019-023-00211-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,5,15]],"date-time":"2023-05-15T14:47:05Z","timestamp":1684162025000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s41019-023-00211-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,15]]},"references-count":63,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2023,6]]}},"alternative-id":["211"],"URL":"https:\/\/doi.org\/10.1007\/s41019-023-00211-0","relation":{},"ISSN":["2364-1185","2364-1541"],"issn-type":[{"value":"2364-1185","type":"print"},{"value":"2364-1541","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,4,15]]},"assertion":[{"value":"29 December 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 March 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 March 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 April 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}