{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,25]],"date-time":"2026-02-25T23:04:04Z","timestamp":1772060644437,"version":"3.50.1"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2026,2,25]],"date-time":"2026-02-25T00:00:00Z","timestamp":1771977600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,25]],"date-time":"2026-02-25T00:00:00Z","timestamp":1771977600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-026-21140-w","type":"journal-article","created":{"date-parts":[[2026,2,25]],"date-time":"2026-02-25T22:33:09Z","timestamp":1772058789000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A comparative evaluation of transformer models for medical abstract classification"],"prefix":"10.1007","volume":"85","author":[{"given":"Mohammad","family":"AnsariShiri","sequence":"first","affiliation":[]},{"given":"Najme","family":"Mansouri","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,25]]},"reference":[{"key":"21140_CR1","doi-asserted-by":"publisher","first-page":"45","DOI":"10.1016\/j.compbiomed.2016.04.016","volume":"74","author":"GS Hern\u00e1ndez-Chan","year":"2016","unstructured":"Hern\u00e1ndez-Chan GS, Ceh-Varela EE, Sanchez-Cervantes JL, Villanueva-Escalante M, Rodr\u00edguez-Gonz\u00e1lez A, P\u00e9rez-Gallardo Y (2016) Collective intelligence in medical diagnosis systems: a case study. Comput Biol Med 74:45\u201353. https:\/\/doi.org\/10.1016\/j.compbiomed.2016.04.016","journal-title":"Comput Biol Med"},{"issue":"1","key":"21140_CR2","first-page":"838","volume":"5","author":"I Parvathi","year":"2014","unstructured":"Parvathi I, Rautaray S (2014) Survey on data mining techniques for the diagnosis of diseases in medical domain. Int J Comput Sci Inf Technol 5(1):838\u2013846","journal-title":"Int J Comput Sci Inf Technol"},{"key":"21140_CR3","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.procs.2016.04.006","volume":"82","author":"S AlMuhaideb","year":"2016","unstructured":"AlMuhaideb S, Menai MEB (2016) An individualized preprocessing for medical data classification. Procedia Comput Sci 82:35\u201342. https:\/\/doi.org\/10.1016\/j.procs.2016.04.006","journal-title":"Procedia Comput Sci"},{"key":"21140_CR4","doi-asserted-by":"publisher","unstructured":"Thong NT, Le Hoang S (2015) HIFCF: an efective hybrid model between picture fuzzy clustering and intuitionistic fuzzy recommender systems for medical diagnosis. In Proceedings of journal of expert systems with applications. 3682\u20133701. https:\/\/doi.org\/10.1016\/j.eswa.2014.12.042","DOI":"10.1016\/j.eswa.2014.12.042"},{"issue":"1","key":"21140_CR5","doi-asserted-by":"publisher","first-page":"57","DOI":"10.1093\/bib\/6.1.57","volume":"6","author":"AM Cohen","year":"2005","unstructured":"Cohen AM, Hersh WR (2005) A survey of current work in biomedical text mining. Brief Bioinform 6(1):57\u201371. https:\/\/doi.org\/10.1093\/bib\/6.1.57","journal-title":"Brief Bioinform"},{"key":"21140_CR6","doi-asserted-by":"publisher","first-page":"345","DOI":"10.1016\/j.neucom.2021.02.069","volume":"443","author":"X Li","year":"2021","unstructured":"Li X, Cui M, Li J, Bai R, Lu Z, Aickelin U (2021) A hybrid medical text classification framework: integrating attentive rule construction and neural network. Neurocomputing 443:345\u2013355. https:\/\/doi.org\/10.1016\/j.neucom.2021.02.069","journal-title":"Neurocomputing"},{"issue":"1","key":"21140_CR7","doi-asserted-by":"publisher","DOI":"10.1186\/s12911-022-01946-y","volume":"22","author":"J Li","year":"2022","unstructured":"Li J, Lin Y, Zhao P, Liu W, Cai L, Sun J, Zhao Lei, Yang Zhenghan, Song Hong, Lv Han, Wang Z (2022) Automatic text classification of actionable radiology reports of tinnitus patients using bidirectional encoder representations from transformer (BERT) and in-domain pre-training (IDPT). BMC Med Inform Decis Mak 22(1):200. https:\/\/doi.org\/10.1186\/s12911-022-01946-y","journal-title":"BMC Med Inform Decis Mak"},{"key":"21140_CR8","doi-asserted-by":"publisher","unstructured":"Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805. https:\/\/doi.org\/10.48550\/arXiv.1810.04805","DOI":"10.48550\/arXiv.1810.04805"},{"key":"21140_CR9","doi-asserted-by":"publisher","first-page":"274","DOI":"10.1016\/j.procs.2015.07.304","volume":"53","author":"J Kim","year":"2015","unstructured":"Kim J, Choe Y, Mueller K (2015) Extracting clinical relations in electronic health records using enriched parse trees. Procedia Comput Sci 53:274\u2013283. https:\/\/doi.org\/10.1016\/j.procs.2015.07.304","journal-title":"Procedia Comput Sci"},{"issue":"1","key":"21140_CR10","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/505282.505283","volume":"34","author":"F Sebastiani","year":"2002","unstructured":"Sebastiani F (2002) Machine learning in automated text categorization. ACM Comput Surv 34(1):1\u201347. https:\/\/doi.org\/10.1145\/505282.505283","journal-title":"ACM Comput Surv"},{"issue":"2","key":"21140_CR11","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0114677","volume":"9","author":"R Cohen","year":"2014","unstructured":"Cohen R, Aviram I, Elhadad M, Elhadad N (2014) Redundancy-aware topic modeling for patient record notes. PLoS One 9(2):e87555. https:\/\/doi.org\/10.1371\/journal.pone.0114677","journal-title":"PLoS One"},{"issue":"5","key":"21140_CR12","doi-asserted-by":"publisher","first-page":"871","DOI":"10.1136\/amiajnl-2014-002694","volume":"21","author":"BJ Marafino","year":"2014","unstructured":"Marafino BJ, Davies JM, Bardach NS, Dean ML, Dudley RA (2014) N-gram support vector machines for scalable procedure and diagnosis classification, with applications to clinical free text data from the intensive care unit. J Am Med Inform Assoc 21(5):871\u2013875. https:\/\/doi.org\/10.1136\/amiajnl-2014-002694","journal-title":"J Am Med Inform Assoc"},{"key":"21140_CR13","doi-asserted-by":"publisher","unstructured":"Johnson R, Zhang T (2015) Semi-supervised convolutional neural networks for text categorization via region embedding.\u00a0Adv Neural Inf Process Syst\u00a028. https:\/\/doi.org\/10.48550\/arXiv.1504.01255","DOI":"10.48550\/arXiv.1504.01255"},{"key":"21140_CR14","doi-asserted-by":"publisher","first-page":"225","DOI":"10.1016\/j.aiopen.2021.08.002","volume":"2","author":"X Han","year":"2021","unstructured":"Han X, Zhang Z, Ding N, Gu Y, Liu X, Huo Y, Qiu Jiezhong, Yao Yuan, Zhang Ao, Zhang Liang, Han Wentao, Huang Minlie, Jin Qin, Lan Yanyan, Liu Yang, Liu Zhiyuan, Lu Zhiwu, Qiu Xipeng, Song Ruihua, Tang Jie, Wen Ji-Rong, Yuan Jinhui, Zhao Wayne Xin, Zhu J (2021) Pre-trained models: past, present and future. AI Open 2:225\u2013250. https:\/\/doi.org\/10.1016\/j.aiopen.2021.08.002","journal-title":"AI Open"},{"key":"21140_CR15","doi-asserted-by":"publisher","unstructured":"Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Stoyanov V (2019) Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692. https:\/\/doi.org\/10.48550\/arXiv.1907.11692","DOI":"10.48550\/arXiv.1907.11692"},{"key":"21140_CR16","doi-asserted-by":"publisher","unstructured":"Lan Z, Chen M., Goodman S, Gimpel K, Sharma P, Soricut R (2019) Albert: a lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942. https:\/\/doi.org\/10.48550\/arXiv.1909.11942","DOI":"10.48550\/arXiv.1909.11942"},{"key":"21140_CR17","doi-asserted-by":"publisher","unstructured":"Sun Y, Wang S, Li Y, Feng S, Chen X, Zhang H, Wu H (2019) Ernie: Enhanced representation through knowledge integration. arXiv preprint arXiv:1904.09223. https:\/\/doi.org\/10.48550\/arXiv.1904.09223","DOI":"10.48550\/arXiv.1904.09223"},{"key":"21140_CR18","doi-asserted-by":"publisher","first-page":"131662","DOI":"10.1109\/ACCESS.2020.3009626","volume":"8","author":"K Mishev","year":"2020","unstructured":"Mishev K, Gjorgjevikj A, Vodenska I, Chitkushev LT, Trajanov D (2020) Evaluation of sentiment analysis in finance: from lexicons to transformers. IEEE Access 8:131662\u2013131682. https:\/\/doi.org\/10.1109\/ACCESS.2020.3009626","journal-title":"IEEE Access"},{"key":"21140_CR19","doi-asserted-by":"publisher","unstructured":"Bataa E, Wu J (2019) An investigation of transfer learning-based sentiment analysis in Japanese. arXiv preprint arXiv:1905.09642. https:\/\/doi.org\/10.48550\/arXiv.1905.09642","DOI":"10.48550\/arXiv.1905.09642"},{"key":"21140_CR20","doi-asserted-by":"publisher","unstructured":"Malte A, Ratadiya P (2019) Evolution of transfer learning in natural language processing. arXiv preprint arXiv:1910.07370. https:\/\/doi.org\/10.48550\/arXiv.1910.07370","DOI":"10.48550\/arXiv.1910.07370"},{"key":"21140_CR21","doi-asserted-by":"publisher","first-page":"176600","DOI":"10.1109\/ACCESS.2019.2953990","volume":"7","author":"S Yu","year":"2019","unstructured":"Yu S, Su J, Luo D (2019) Improving bert-based text classification with auxiliary sentence and domain knowledge. IEEE Access 7:176600\u2013176612. https:\/\/doi.org\/10.1109\/ACCESS.2019.2953990","journal-title":"IEEE Access"},{"key":"21140_CR22","doi-asserted-by":"publisher","unstructured":"Huang K, Altosaar J, Ranganath R (2019) Clinicalbert: modeling clinical notes and predicting hospital readmission. arXiv preprint arXiv:1904.05342. https:\/\/doi.org\/10.48550\/arXiv.1904.05342","DOI":"10.48550\/arXiv.1904.05342"},{"issue":"4","key":"21140_CR23","doi-asserted-by":"publisher","first-page":"1234","DOI":"10.1093\/bioinformatics\/btz682","volume":"36","author":"J Lee","year":"2020","unstructured":"Lee J, Yoon W, Kim S, Kim D, Kim S, So CH, Kang J (2020) BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36(4):1234\u20131240. https:\/\/doi.org\/10.1093\/bioinformatics\/btz682","journal-title":"Bioinformatics"},{"key":"21140_CR24","doi-asserted-by":"publisher","DOI":"10.1155\/2022\/3498123","author":"R Qasim","year":"2022","unstructured":"Qasim R, Bangyal WH, Alqarni MA, Ali Almazroi A (2022) A fine-tuned BERT-based transfer learning approach for text classification. J Healthc Eng. https:\/\/doi.org\/10.1155\/2022\/3498123","journal-title":"J Healthc Eng"},{"issue":"8","key":"21140_CR25","doi-asserted-by":"publisher","first-page":"11765","DOI":"10.1007\/s11042-020-10183-2","volume":"80","author":"RK Kaliyar","year":"2021","unstructured":"Kaliyar RK, Goswami A, Narang P (2021) FakeBERT: fake news detection in social media with a BERT-based deep learning approach. Multimed Tools Appl 80(8):11765\u201311788. https:\/\/doi.org\/10.1007\/s11042-020-10183-2","journal-title":"Multimed Tools Appl"},{"key":"21140_CR26","doi-asserted-by":"publisher","unstructured":"Guo T, Gao H (2019) Content enhanced bert-based text-to-sql generation. arXiv preprint arXiv:1910.07179. https:\/\/doi.org\/10.48550\/arXiv.1910.07179","DOI":"10.48550\/arXiv.1910.07179"},{"key":"21140_CR27","doi-asserted-by":"publisher","unstructured":"Sun C, Huang L, Qiu X. (2019) Utilizing BERT for aspect-based sentiment analysis via constructing auxiliary sentence. arXiv preprint arXiv:1903.09588. https:\/\/doi.org\/10.48550\/arXiv.1903.09588","DOI":"10.48550\/arXiv.1903.09588"},{"key":"21140_CR28","doi-asserted-by":"publisher","unstructured":"Nikolov A, Radivchev V (2019) Nikolov-radivchev at semeval-2019 task 6: Offensive tweet classification with bert and ensembles. In Proceedings of the 13th international workshop on semantic evaluation.\u00a0pp. 691\u2013695. https:\/\/doi.org\/10.18653\/v1\/S19-2123","DOI":"10.18653\/v1\/S19-2123"},{"key":"21140_CR29","unstructured":"Goldberg Y (2019) Assessing BERT\u2019s syntactic abilities. arXiv preprint arXiv:1901.05287."},{"key":"21140_CR30","doi-asserted-by":"publisher","unstructured":"Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Stoyanov V (2019) Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692. https:\/\/doi.org\/10.48550\/arXiv.1907.1169","DOI":"10.48550\/arXiv.1907.1169"},{"key":"21140_CR31","doi-asserted-by":"publisher","unstructured":"Warstadt A, Zhang Y, Li HS, Liu H, Bowman SR (2020) Learning which features matter: RoBERTa acquires a preference for linguistic generalizations (eventually). arXiv preprint arXiv:2010.05358. https:\/\/doi.org\/10.48550\/arXiv.2010.05358","DOI":"10.48550\/arXiv.2010.05358"},{"key":"21140_CR32","doi-asserted-by":"publisher","unstructured":"Murarka A, Radhakrishnan B, Ravichandran S (2020) Detection and classification of mental illnesses on social media using RoBERTa. arXiv preprint arXiv:2011.11226. https:\/\/doi.org\/10.48550\/arXiv.2011.11226","DOI":"10.48550\/arXiv.2011.11226"},{"key":"21140_CR33","doi-asserted-by":"publisher","first-page":"2437","DOI":"10.1109\/TASLP.2021.3077119","volume":"29","author":"B Kim","year":"2021","unstructured":"Kim B, Seo J, Koo MW (2021) Randomly wired network based on RoBERTa and dialog history attention for response selection. IEEE ACM Trans Audio Speech Lang Process 29:2437\u20132442. https:\/\/doi.org\/10.1109\/TASLP.2021.3077119","journal-title":"IEEE ACM Trans Audio Speech Lang Process"},{"key":"21140_CR34","unstructured":"Casola S, Lavelli A (2020) FBK@ SMM4H2020: RoBERTa for detecting medications on Twitter. In Proceedings of the Fifth Social Media Mining for Health Applications Workshop & Shared Task (pp. 101\u2013103)."},{"key":"21140_CR35","doi-asserted-by":"publisher","unstructured":"Delobelle P, Winters T, Berendt B (2020) Robbert: a dutch roberta-based language model. arXiv preprint arXiv:2001.06286. https:\/\/doi.org\/10.48550\/arXiv.2001.06286","DOI":"10.48550\/arXiv.2001.06286"},{"key":"21140_CR36","doi-asserted-by":"publisher","unstructured":"Li B, Rudzicz F (2021) TorontoCL at CMCL 2021 shared task RoBERTa with multi-stage fine-tuning for eye-tracking prediction. arXiv preprint arXiv:2104.07244. https:\/\/doi.org\/10.48550\/arXiv.2104.07244","DOI":"10.48550\/arXiv.2104.07244"},{"key":"21140_CR37","doi-asserted-by":"publisher","unstructured":"Sanh V, Debut L, Chaumond J, Wolf T (2019) DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108. https:\/\/doi.org\/10.48550\/arXiv.1910.01108","DOI":"10.48550\/arXiv.1910.01108"},{"key":"21140_CR38","unstructured":"B\u00fcy\u00fck\u00f6z B, H\u00fcrriyeto\u011flu A, \u00d6zg\u00fcr A (2020) Analyzing ELMo and DistilBERT on socio-political news classification. In Proceedings of the Workshop on Automated Extraction of Socio-political Events from News 2020 (pp. 9\u201318)."},{"key":"21140_CR39","doi-asserted-by":"publisher","unstructured":"Stali\u016bnait\u0117 I, Iacobacci I (2020). Compositional and lexical semantics in RoBERTa, BERT and DistilBERT: a case study on CoQA. arXiv preprint arXiv:2009.08257. https:\/\/doi.org\/10.48550\/arXiv.2009.08257","DOI":"10.48550\/arXiv.2009.08257"},{"key":"21140_CR40","doi-asserted-by":"publisher","unstructured":"Abadeer M (2020) Assessment of DistilBERT performance on named entity recognition task for the detection of protected health information and medical concepts. In Proceedings of the 3rd clinical natural language processing workshop. 158\u2013167. https:\/\/doi.org\/10.18653\/v1\/2020.clinicalnlp-1.18","DOI":"10.18653\/v1\/2020.clinicalnlp-1.18"},{"key":"21140_CR41","doi-asserted-by":"publisher","unstructured":"Mozafari J, Fatemi A, Moradi P (2020) A method for answer selection using DistilBERT and important words. In 2020 6th International Conference on Web Research (ICWR). pp. 72\u201376. IEEE. https:\/\/doi.org\/10.1109\/ICWR49608.2020.9122302","DOI":"10.1109\/ICWR49608.2020.9122302"},{"key":"21140_CR42","doi-asserted-by":"publisher","unstructured":"Lan Z, Chen M, Goodman S, Gimpel K, Sharma P, Soricut R (2019) Albert: a lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942. https:\/\/doi.org\/10.48550\/arXiv.1909.11942","DOI":"10.48550\/arXiv.1909.11942"},{"key":"21140_CR43","doi-asserted-by":"publisher","unstructured":"Gregory H, Li S, Mohammadi P, Tarn N, Draelos R, Rudin C (2020) A transformer approach to contextual sarcasm detection in twitter. In Proceedings of the second workshop on figurative language processing.\u00a0270\u2013275. https:\/\/doi.org\/10.18653\/v1\/2020.figlang-1.37","DOI":"10.18653\/v1\/2020.figlang-1.37"},{"key":"21140_CR44","doi-asserted-by":"publisher","unstructured":"Vijjali R, Potluri P, Kumar S, Teki S (2020). Two stage transformer model for COVID-19 fake news detection and fact checking. arXiv preprint arXiv:2011.13253. https:\/\/doi.org\/10.48550\/arXiv.2011.13253","DOI":"10.48550\/arXiv.2011.13253"},{"key":"21140_CR45","doi-asserted-by":"publisher","unstructured":"Liu Y, Ott,M, Goyal N, Du J, Joshi M, Chen D, Stoyanov V (2019) Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692. https:\/\/doi.org\/10.48550\/arXiv.1907.11692","DOI":"10.48550\/arXiv.1907.11692"},{"key":"21140_CR46","unstructured":"Zhao Y, Tao X (2021) ZYJ123@ DravidianLangTech-EACL2021: Offensive language identification based on XLM-RoBERTa with DPCNN. In Proceedings of the first workshop on speech and language technologies for dravidian languages.\u00a0216\u2013221."},{"key":"21140_CR47","unstructured":"Ou X, Li H (2020) YNU@ dravidian-codemix-FIRE2020: XLM-RoBERTa for multi-language Sentiment Analysis. In FIRE (Working Notes).\u00a0560\u2013565."},{"key":"21140_CR48","doi-asserted-by":"publisher","unstructured":"Xie S, Ma J, Yang H, Jiang L, Mo, Y, Shen J (2021) PALI at SemEval-2021 task 2: fine-tune XLM-RoBERTa for word in context disambiguation. arXiv preprint arXiv:2104.10375.https:\/\/doi.org\/10.48550\/arXiv.2104.10375","DOI":"10.48550\/arXiv.2104.10375"},{"key":"21140_CR49","doi-asserted-by":"publisher","unstructured":"Clark K, Luong MT, Le QV, Manning CD (2020) Electra: pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555. https:\/\/doi.org\/10.48550\/arXiv.2003.10555","DOI":"10.48550\/arXiv.2003.10555"},{"key":"21140_CR50","unstructured":"Das KA, Baruah A, Barbhuiya FA, Dey K (2020) Ensemble of ELECTRA for profiling fake news spreaders. In CLEF (Working Notes)."},{"key":"21140_CR51","doi-asserted-by":"publisher","unstructured":"Lewis M, Liu Y, Goyal N, Ghazvininejad M, Mohamed A, Levy O, Zettlemoyer L (2019) Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461. https:\/\/doi.org\/10.48550\/arXiv.1910.13461","DOI":"10.48550\/arXiv.1910.13461"},{"key":"21140_CR52","unstructured":"Mustar A, Lamprier S, Piwowarski B (2020) Using BERT and BART for query suggestion. In Joint Conference of the Information Retrieval Communities in Europe. 2621. CEUR-WS. org."},{"key":"21140_CR53","doi-asserted-by":"publisher","unstructured":"Xing Y, Shi Z, Meng Z, Lakemeyer G, Ma Y, Wattenhofer R (2021) Km-bart: Knowledge enhanced multimodal bart for visual commonsense generation. arXiv preprint arXiv:2101.00419. https:\/\/doi.org\/10.48550\/arXiv.2101.00419","DOI":"10.48550\/arXiv.2101.00419"},{"key":"21140_CR54","doi-asserted-by":"crossref","unstructured":"Alsentzer E, Murphy JR, Boag W, Weng WH, Jindi D, Naumann T, McDermott M (2019) Publicly available clinical BERT embeddings. arXiv preprint arXiv:1904.03323.","DOI":"10.18653\/v1\/W19-1909"},{"key":"21140_CR55","doi-asserted-by":"crossref","unstructured":"Beltagy I, Lo K, Cohan A (2019) SciBERT: a pretrained language model for scientific text. arXiv preprint arXiv:1903.10676.","DOI":"10.18653\/v1\/D19-1371"},{"key":"21140_CR56","doi-asserted-by":"publisher","first-page":"172","DOI":"10.1016\/j.ins.2020.03.006","volume":"525","author":"N Shanavas","year":"2020","unstructured":"Shanavas N, Wang H, Lin Z, Hawe G (2020) Ontology-based enriched concept graphs for medical document classification. Inf Sci 525:172\u2013181. https:\/\/doi.org\/10.1016\/j.ins.2020.03.006","journal-title":"Inf Sci"},{"key":"21140_CR57","doi-asserted-by":"publisher","unstructured":"Zhang Y, Yu X, Cui Z, Wu S, Wen Z, Wang L (2020) Every document owns its structure: Inductive text classification via graph neural networks. https:\/\/doi.org\/10.48550\/arXiv.2004.13826","DOI":"10.48550\/arXiv.2004.13826"},{"key":"21140_CR58","doi-asserted-by":"publisher","unstructured":"Yao L, Mao C, Luo Y (2019) Graph convolutional networks for text classification. In Proceedings of the AAAI conference on artificial intelligence.\u00a033(01):7370\u20137377. https:\/\/doi.org\/10.1609\/aaai.v33i01.33017370","DOI":"10.1609\/aaai.v33i01.33017370"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-026-21140-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-026-21140-w","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-026-21140-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,25]],"date-time":"2026-02-25T22:33:16Z","timestamp":1772058796000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-026-21140-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,25]]},"references-count":58,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2026,3]]}},"alternative-id":["21140"],"URL":"https:\/\/doi.org\/10.1007\/s11042-026-21140-w","relation":{},"ISSN":["1573-7721"],"issn-type":[{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,25]]},"assertion":[{"value":"16 February 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 June 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 December 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 February 2026","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval and consent to participate"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"The authors declare that they have no competing interests.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"183"}}