{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,21]],"date-time":"2026-02-21T13:09:13Z","timestamp":1771679353054,"version":"3.50.1"},"reference-count":63,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2022,8,18]],"date-time":"2022-08-18T00:00:00Z","timestamp":1660780800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,8,18]],"date-time":"2022-08-18T00:00:00Z","timestamp":1660780800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Process Lett"],"published-print":{"date-parts":[[2023,6]]},"DOI":"10.1007\/s11063-022-10990-8","type":"journal-article","created":{"date-parts":[[2022,8,18]],"date-time":"2022-08-18T21:02:45Z","timestamp":1660856565000},"page":"2911-2933","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":29,"title":["Evaluating Various Tokenizers for Arabic Text Classification"],"prefix":"10.1007","volume":"55","author":[{"given":"Zaid","family":"Alyafeai","sequence":"first","affiliation":[]},{"given":"Maged S.","family":"Al-shaibani","sequence":"additional","affiliation":[]},{"given":"Mustafa","family":"Ghaleb","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8311-1731","authenticated-orcid":false,"given":"Irfan","family":"Ahmad","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,8,18]]},"reference":[{"issue":"2","key":"10990_CR1","first-page":"103","volume":"6","author":"G Abandah","year":"2020","unstructured":"Abandah G, Abdel-Karim A (2020) Accurate and fast recurrent neural network solution for the automatic diacritization of arabic text. Jordanian J Comput Inf Tech 6(2):103\u2013121","journal-title":"Jordanian J Comput Inf Tech"},{"issue":"6","key":"10990_CR2","first-page":"3775","volume":"34","author":"A Abandah Gheith","year":"2020","unstructured":"Abandah Gheith A, Khedher Mohammed Z, Abdel-Majeed Mohammad R, Mansour Hamdi M, Hulliel Salma F, Bisharat Lara M (2020) Classifying and diacritizing arabic poems using deep recurrent neural networks. J King Saud Univ Comput Inf Sci 34(6):3775\u20133788","journal-title":"J King Saud Univ Comput Inf Sci"},{"key":"10990_CR3","doi-asserted-by":"crossref","unstructured":"Abdelali Ahmed, Darwish Kareem, Durrani Nadir, Mubarak Hamdy (2016) Farasa: A fast and furious segmenter for arabic. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Demonstrations, pp 11\u201316","DOI":"10.18653\/v1\/N16-3003"},{"key":"10990_CR4","unstructured":"Abdelali Ahmed, Hassan Sabit, Mubarak Hamdy, Darwish Kareem, Samih Younes (2021) Pre-training bert on arabic tweets: Practical considerations. arXiv preprint arXiv:2102.10684"},{"key":"10990_CR5","doi-asserted-by":"crossref","unstructured":"Abdul-Mageed Muhammad, Elmadany AbdelRahim, Nagoudi El\u00a0Moatez\u00a0Billah (2020) Arbert & marbert: deep bidirectional transformers for arabic. arXiv preprint arXiv:2101.01785","DOI":"10.18653\/v1\/2021.acl-long.551"},{"key":"10990_CR6","doi-asserted-by":"publisher","unstructured":"Farha Ibrahim Abu, Magdy Walid (2021) A comparative study of effective approaches for arabic sentiment analysis. Information Processing & Management 58(2):102438. ISSN 0306-4573. https:\/\/doi.org\/10.1016\/j.ipm.2020.102438. URL http:\/\/www.sciencedirect.com\/science\/article\/pii\/S0306457320309316","DOI":"10.1016\/j.ipm.2020.102438"},{"issue":"2","key":"10990_CR7","doi-asserted-by":"publisher","first-page":"320","DOI":"10.1016\/j.ipm.2018.07.006","volume":"56","author":"M Al-Ayyoub","year":"2019","unstructured":"Al-Ayyoub M, Khamaiseh AA, Jararweh Y, Al-Kabi MN (2019) A comprehensive survey of arabic sentiment analysis. Inf process manag 56(2):320\u2013342","journal-title":"Inf process manag"},{"issue":"3","key":"10990_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3060620","volume":"50","author":"BM Al-Helali","year":"2017","unstructured":"Al-Helali BM, Mahmoud SA (2017) Arabic online handwriting recognition (aohr) a survey. ACM Comput Surveys (CSUR) 50(3):1\u201335","journal-title":"ACM Comput Surveys (CSUR)"},{"key":"10990_CR9","first-page":"3159","volume":"33","author":"R Al-Rfou","year":"2019","unstructured":"Al-Rfou R, Choe D, Constant N, Guo M, Jones L (2019) Character-level language modeling with deeper self-attention. Proc AAAI Conf Artif Intell 33:3159\u20133166","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"10990_CR10","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.patrec.2020.05.028","volume":"136","author":"S Al-shaibani Maged","year":"2020","unstructured":"Al-shaibani Maged S, Zaid Alyafeai, Irfan Ahmad (2020) Meter classification of arabic poems using deep bidirectional recurrent neural networks. Pattern Recognition Letters 136:1\u20137","journal-title":"Pattern Recognition Letters"},{"key":"10990_CR11","doi-asserted-by":"publisher","DOI":"10.1016\/j.dib.2020.106497","volume":"33","author":"MS Al-Shaibani","year":"2020","unstructured":"Al-Shaibani MS, Alyafeai Z, Ahmad I (2020) Metrec: A dataset for meter classification of arabic poetry. Data Brief 33:106497","journal-title":"Data Brief"},{"key":"10990_CR12","unstructured":"Alkaoud Mohamed, Syed Mairaj (2020) On the importance of tokenization in arabic embedding models. In: Proceedings of the Fifth Arabic Natural Language Processing Workshop, pp 119\u2013129"},{"key":"10990_CR13","doi-asserted-by":"crossref","unstructured":"Alomari Khaled\u00a0Mohammad, ElSherif Hatem\u00a0M, Shaalan Khaled (2017) Arabic tweets sentimental analysis using machine learning. In: International Conference on Industrial, Engineering and Other Applications of Applied Intelligent Systems, pp 602\u2013610. Springer","DOI":"10.1007\/978-3-319-60042-0_66"},{"key":"10990_CR14","unstructured":"Aly Mohamed, Atiya Amir (2013) Labr: A large scale arabic book reviews dataset. In: Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp 494\u2013498"},{"key":"10990_CR15","unstructured":"Antoun Wissam, Baly Fady, Hajj Hazem (2020) Arabert: Transformer-based model for arabic language understanding. arXiv preprint arXiv:2003.00104"},{"key":"10990_CR16","doi-asserted-by":"crossref","unstructured":"Atallah AL-Shatnawi, Omar Khairuddin (2009) A comparative study between methods of arabic baseline detection. In: 2009 International Conference on Electrical Engineering and Informatics, volume\u00a01, pp 73\u201377. IEEE","DOI":"10.1109\/ICEEI.2009.5254814"},{"issue":"3","key":"10990_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3295662","volume":"18","author":"G Badaro","year":"2019","unstructured":"Badaro G, Baly R, Hajj H, El-Hajj W, Shaban KB, Habash N, Al-Sallab A, Hamdi A (2019) A survey of opinion mining in arabic: a comprehensive system perspective covering challenges and advances in tools, resources, models, applications, and visualizations. ACM Trans Asian Low-Resource Lang Inf Process (TALLIP) 18(3):1\u201352","journal-title":"ACM Trans Asian Low-Resource Lang Inf Process (TALLIP)"},{"key":"10990_CR18","unstructured":"mohamed BINIZ. Dataset for arabic classification. (2018)"},{"key":"10990_CR19","doi-asserted-by":"publisher","first-page":"135","DOI":"10.1162\/tacl_a_00051","volume":"5","author":"P Bojanowski","year":"2017","unstructured":"Bojanowski P, Grave E, Joulin A, Mikolov T (2017) Enriching word vectors with subword information. Trans Assoc Comput Linguist 5:135\u2013146 (ISSN 2307-387X)","journal-title":"Trans Assoc Comput Linguist"},{"key":"10990_CR20","doi-asserted-by":"crossref","unstructured":"Bostrom Kaj, Durrett Greg (2020) Byte pair encoding is suboptimal for language model pretraining. arXiv preprint arXiv:2004.03720","DOI":"10.18653\/v1\/2020.findings-emnlp.414"},{"key":"10990_CR21","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A et al (2020) Language models are few-shot learners. Adv Neural Inf Process Syst 33:1877\u20131901","journal-title":"Adv Neural Inf Process Syst"},{"key":"10990_CR22","doi-asserted-by":"crossref","unstructured":"Chitnis Rohan, DeNero John (2015) Variable-length word encodings for neural translation models. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp 2088\u20132093","DOI":"10.18653\/v1\/D15-1249"},{"key":"10990_CR23","doi-asserted-by":"crossref","unstructured":"Cho Kyunghyun, Van\u00a0Merri\u00ebnboer Bart, Bahdanau Dzmitry, Bengio Yoshua (2014) On the properties of neural machine translation: Encoder-decoder approaches. arXiv preprint arXiv:1409.1259","DOI":"10.3115\/v1\/W14-4012"},{"key":"10990_CR24","unstructured":"Chollet Fran\u00e7ois et\u00a0al (2015) Keras. https:\/\/github.com\/fchollet\/keras"},{"key":"10990_CR25","doi-asserted-by":"publisher","first-page":"3504","DOI":"10.1109\/TASLP.2021.3124365","volume":"29","author":"Y Cui","year":"2021","unstructured":"Cui Y, Che W, Liu T, Qin B, Yang Z (2021) Pre-training with whole word masking for chinese bert. IEEE\/ACM Trans Audio, Speech, Lang Process 29:3504\u20133514","journal-title":"IEEE\/ACM Trans Audio, Speech, Lang Process"},{"key":"10990_CR26","unstructured":"Devlin Jacob, Chang Ming-Wei, Lee Kenton, Toutanova Kristina (2018) Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805"},{"issue":"1","key":"10990_CR27","first-page":"1","volume":"19","author":"C Ding","year":"2019","unstructured":"Ding C, Aye HTZ, Pa WP, Nwet KT, Soe KM, Utiyama M, Sumita E (2019) Towards burmese (myanmar) morphological analysis: Syllable-based tokenization and part-of-speech tagging. ACM Trans Asian Low-Resource Lang Inf Process (TALLIP) 19(1):1\u201334","journal-title":"ACM Trans Asian Low-Resource Lang Inf Process (TALLIP)"},{"key":"10990_CR28","unstructured":"El-Khair Ibrahim\u00a0Abu (2016) 1.5 billion words arabic corpus. arXiv preprint arXiv:1611.04033"},{"key":"10990_CR29","doi-asserted-by":"crossref","unstructured":"ElJundi Obeida, Antoun Wissam, Droubi Nour El, Hajj Hazem, El-Hajj Wassim, Shaban Khaled (2019) hulmona: The universal language model in arabic. In: Proceedings of the Fourth Arabic Natural Language Processing Workshop, pp 68\u201377","DOI":"10.18653\/v1\/W19-4608"},{"issue":"1","key":"10990_CR30","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1007\/s13278-019-0602-x","volume":"9","author":"I Guellil","year":"2019","unstructured":"Guellil I, Azouaou F, Mendoza M (2019) Arabic sentiment analysis: studies, resources, and tools. Soc Netw Anal Min 9(1):56","journal-title":"Soc Netw Anal Min"},{"key":"10990_CR31","doi-asserted-by":"crossref","unstructured":"Howard Jeremy, Ruder Sebastian (2018) Universal language model fine-tuning for text classification. arXiv preprint arXiv:1801.06146","DOI":"10.18653\/v1\/P18-1031"},{"key":"10990_CR32","unstructured":"Jasim Mahdi\u00a0Nsaif (2020) Arabic optical characters recognition by neural network based arabic unicode"},{"key":"10990_CR33","doi-asserted-by":"crossref","unstructured":"Kudo Taku (2018) Subword regularization: Improving neural network translation models with multiple subword candidates. arXiv preprint arXiv:1804.10959,","DOI":"10.18653\/v1\/P18-1007"},{"key":"10990_CR34","doi-asserted-by":"crossref","unstructured":"Kudo Taku, Richardson John (2018) Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. arXiv preprint arXiv:1808.06226","DOI":"10.18653\/v1\/D18-2012"},{"key":"10990_CR35","doi-asserted-by":"crossref","unstructured":"Kunchukuttan Anoop, Bhattacharyya Pushpak (2016) Orthographic syllable as basic unit for smt between related languages. arXiv preprint arXiv:1610.00634","DOI":"10.18653\/v1\/D16-1196"},{"key":"10990_CR36","unstructured":"Kuratov Yuri, Arkhipov Mikhail (2019) Adaptation of deep bidirectional multilingual transformers for russian language. arXiv preprint arXiv:1905.07213"},{"key":"10990_CR37","doi-asserted-by":"crossref","unstructured":"Lee Sangah, Shin Hyopil (2021) The korean morphologically tight-fitting tokenizer for noisy user-generated texts. In: Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021), pp 410\u2013416","DOI":"10.18653\/v1\/2021.wnut-1.45"},{"key":"10990_CR38","unstructured":"Liu Yinhan, Ott Myle, Goyal Naman, Du Jingfei, Joshi Mandar, Chen Danqi, Levy Omer, Lewis Mike, Zettlemoyer Luke, Stoyanov Veselin (2019) Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692"},{"key":"10990_CR39","doi-asserted-by":"crossref","unstructured":"Ma Wentao, Cui Yiming, Si Chenglei, Liu Ting, Wang Shijin, Hu Guoping (2020) Charbert: Character-aware pre-trained language model. arXiv preprint arXiv:2011.01513","DOI":"10.18653\/v1\/2020.coling-main.4"},{"key":"10990_CR40","doi-asserted-by":"crossref","unstructured":"Martin Louis, Muller Benjamin, Su\u00e1rez Pedro Javier\u00a0Ortiz, Dupont Yoann, Romary Laurent, de\u00a0La\u00a0Clergerie \u00c9ric\u00a0Villemonte, Seddah Djam\u00e9, Sagot Beno\u00eet (2019) Camembert: a tasty french language model. arXiv preprint arXiv:1911.03894","DOI":"10.18653\/v1\/2020.acl-main.645"},{"key":"10990_CR41","unstructured":"Mikolov Tom\u00e1\u0161, Sutskever Ilya, Deoras Anoop, Le Hai-Son, Kombrink Stefan, Cernocky Jan (2012) Subword language modeling with neural networks. preprint (http:\/\/www.fit.vutbr.cz\/mikolov\/rnnlm\/char.pdf), 8:67"},{"key":"10990_CR42","unstructured":"Mikolov Tomas, Chen Kai, Corrado Greg, Dean Jeffrey (2013) Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781"},{"key":"10990_CR43","doi-asserted-by":"publisher","unstructured":"Mubarak Hamdy, Abdelali Ahmed, Sajjad Hassan, Samih Younes, Darwish Kareem (June 2019) Highly effective Arabic diacritization using sequence to sequence modeling. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp 2390\u20132395, Minneapolis, Minnesota. Association for Computational Linguistics. https:\/\/doi.org\/10.18653\/v1\/N19-1248","DOI":"10.18653\/v1\/N19-1248"},{"key":"10990_CR44","unstructured":"Oudah Mai, Almahairi Amjad, Habash Nizar (2019) The impact of preprocessing on arabic-english statistical and neural machine translation. arXiv preprint arXiv:1906.11751"},{"key":"10990_CR45","unstructured":"Pasha Arfath, Al-Badrashiny Mohamed, Diab Mona\u00a0T, Kholy Ahmed El, Eskander Ramy, Habash Nizar, Pooleery Manoj (2014) Owen Rambow, and Ryan Roth. Madamira: A fast, comprehensive tool for morphological analysis and disambiguation of arabic. In: LREC, volume\u00a014, pp 1094\u20131101"},{"key":"10990_CR46","doi-asserted-by":"crossref","unstructured":"Pennington Jeffrey, Socher Richard, Manning Christopher\u00a0D (2014) Glove: Global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 1532\u20131543","DOI":"10.3115\/v1\/D14-1162"},{"key":"10990_CR47","unstructured":"Radford Alec, Narasimhan Karthik, Salimans Tim, Sutskever Ilya (2018) Improving language understanding by generative pre-training"},{"issue":"8","key":"10990_CR48","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford A, Jeffrey W, Child R, Luan D, Amodei D, Sutskever I (2019) Language models are unsupervised multitask learners. OpenAI Blog 1(8):9","journal-title":"OpenAI Blog"},{"key":"10990_CR49","unstructured":"Raffel Colin, Shazeer Noam, Roberts Adam, Lee Katherine, Narang Sharan, Matena Michael, Zhou Yanqi, Li Wei, Liu Peter\u00a0J (2019) Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv preprint arXiv:1910.10683"},{"key":"10990_CR50","doi-asserted-by":"crossref","unstructured":"Sajjad Hassan, Dalvi Fahim, Durrani Nadir, Abdelali Ahmed, Belinkov Yonatan, Vogel Stephan (2017) Challenging language-dependent segmentation for arabic: An application to machine translation and part-of-speech tagging. arXiv preprint arXiv:1709.00616","DOI":"10.18653\/v1\/P17-2095"},{"key":"10990_CR51","doi-asserted-by":"crossref","unstructured":"Schuster Mike, Nakajima Kaisuke (2012) Japanese and korean voice search. In: 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 5149\u20135152. IEEE","DOI":"10.1109\/ICASSP.2012.6289079"},{"key":"10990_CR52","doi-asserted-by":"crossref","unstructured":"Sennrich Rico, Haddow Barry, Birch Alexandra (2015) Neural machine translation of rare words with subword units. arXiv preprint arXiv:1508.07909","DOI":"10.18653\/v1\/P16-1162"},{"key":"10990_CR53","unstructured":"Shapiro Pamela, Duh Kevin (2018a) Bpe and charcnns for translation of morphology: A cross-lingual comparison and analysis. arXiv preprint arXiv:1809.01301"},{"key":"10990_CR54","doi-asserted-by":"crossref","unstructured":"Shapiro Pamela, Duh Kevin (2018b) Morphological word embeddings for arabic neural machine translation in low-resource settings. In: Proceedings of the Second Workshop on Subword\/Character LEvel Models, pp 1\u201311","DOI":"10.18653\/v1\/W18-1201"},{"key":"10990_CR55","unstructured":"Si Chenglei, Zhang Zhengyan, Chen Yingfa, Qi Fanchao, Wang Xiaozhi, Liu Zhiyuan, Sun Maosong (2021) Shuowen-jiezi: Linguistically informed tokenizers for chinese language model pretraining. arXiv preprint arXiv:2106.00400"},{"key":"10990_CR56","doi-asserted-by":"crossref","unstructured":"Smit Peter, Virpioja Sami, Gr\u00f6nroos Stig-Arne, Kurimo Mikko (2014) Morfessor 2.0: Toolkit for statistical morphological segmentation. pp\u00a04. Aalto University. URL http:\/\/urn.fi\/URN:NBN:fi:aalto-201409292677","DOI":"10.3115\/v1\/E14-2006"},{"key":"10990_CR57","doi-asserted-by":"publisher","first-page":"256","DOI":"10.1016\/j.procs.2017.10.117","volume":"117","author":"AB Soliman","year":"2017","unstructured":"Soliman AB, Eissa K, El-Beltagy SR (2017) Aravec: A set of arabic word embedding models for use in arabic nlp. Procedia Comput Sci 117:256\u2013265","journal-title":"Procedia Comput Sci"},{"key":"10990_CR58","unstructured":"Takaoka Kazuma, Hisamoto Sorami, Kawahara Noriko, Sakamoto Miho, Uchida Yoshitaka, Matsumoto Yuji (2018) Sudachi: A japanese tokenizer for business. In: Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018)"},{"key":"10990_CR59","unstructured":"Tay Yi, Tran Vinh\u00a0Q, Ruder Sebastian, Gupta Jai, Chung Hyung\u00a0Won, Bahri Dara, Qin Zhen, Baumgartner Simon, Yu Cong, Metzler Donald (2021) Charformer: Fast character transformers via gradient-based subword tokenization. arXiv preprint arXiv:2106.12672"},{"key":"10990_CR60","doi-asserted-by":"crossref","unstructured":"Wang Xinyi, Ruder Sebastian, Neubig Graham (2021) Multi-view subword regularization. arXiv preprint arXiv:2103.08490","DOI":"10.18653\/v1\/2021.naacl-main.40"},{"key":"10990_CR61","unstructured":"Wu Yonghui, Schuster Mike, Chen Zhifeng, Le Quoc\u00a0V, Norouzi Mohammad, Macherey Wolfgang, Krikun Maxim, Cao Yuan, Gao Qin, Macherey Klaus et\u00a0al (2016) Google\u2019s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144"},{"key":"10990_CR62","doi-asserted-by":"crossref","unstructured":"Xue Linting, Barua Aditya, Constant Noah, Al-Rfou Rami, Narang Sharan, Kale Mihir, Roberts Adam, Raffel Colin (2021) Byt5: Towards a token-free future with pre-trained byte-to-byte models. arXiv preprint arXiv:2105.13626","DOI":"10.1162\/tacl_a_00461"},{"key":"10990_CR63","unstructured":"Yousef Waleed\u00a0A, Ibrahime Omar\u00a0M, Madbouly Taha\u00a0M, Mahmoud Moustafa\u00a0A (2019) Learning meters of arabic and english poems with recurrent neural networks: a step forward for language understanding and synthesis. arXiv preprint arXiv:1905.05700"}],"container-title":["Neural Processing Letters"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-022-10990-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11063-022-10990-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-022-10990-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,8]],"date-time":"2023-07-08T12:11:45Z","timestamp":1688818305000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11063-022-10990-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,8,18]]},"references-count":63,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2023,6]]}},"alternative-id":["10990"],"URL":"https:\/\/doi.org\/10.1007\/s11063-022-10990-8","relation":{},"ISSN":["1370-4621","1573-773X"],"issn-type":[{"value":"1370-4621","type":"print"},{"value":"1573-773X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,8,18]]},"assertion":[{"value":"28 July 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 August 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"None.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}