{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,20]],"date-time":"2026-04-20T10:25:02Z","timestamp":1776680702635,"version":"3.51.2"},"reference-count":45,"publisher":"Tech Science Press","issue":"1","license":[{"start":{"date-parts":[[2025,3,30]],"date-time":"2025-03-30T00:00:00Z","timestamp":1743292800000},"content-version":"vor","delay-in-days":88,"URL":"https:\/\/doi.org\/10.32604\/TSP-CROSSMARKPOLICY"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["CMC"],"published-print":{"date-parts":[[2025]]},"DOI":"10.32604\/cmc.2025.061527","type":"journal-article","created":{"date-parts":[[2025,3,18]],"date-time":"2025-03-18T04:12:18Z","timestamp":1742271138000},"page":"201-217","update-policy":"https:\/\/doi.org\/10.32604\/tsp-crossmarkpolicy","source":"Crossref","is-referenced-by-count":3,"title":["Multilingual Text Summarization in Healthcare Using Pre-Trained Transformer-Based Language Models"],"prefix":"10.32604","volume":"83","author":[{"given":"Josua","family":"K\u00e4ser","sequence":"first","affiliation":[]},{"given":"Thomas","family":"Nagy","sequence":"additional","affiliation":[]},{"given":"Patrick","family":"Stirnemann","sequence":"additional","affiliation":[]},{"given":"Thomas","family":"Hanne","sequence":"additional","affiliation":[]}],"member":"17807","published-online":{"date-parts":[[2025]]},"reference":[{"key":"ref1","unstructured":"Armengol-Estap\u00e9 J, de Gibert Bonet O, Melero M. On the multilingual capabilities of very large-scale English language models. arXiv:2108.13349. 2021."},{"key":"ref2","doi-asserted-by":"crossref","first-page":"101979","DOI":"10.1016\/j.datak.2022.101979","article-title":"Extraction and analysis of text patterns from NSFW adult content in Reddit","volume":"138","author":"Cauteruccio","year":"2022","journal-title":"Data Knowl Eng"},{"key":"ref3","doi-asserted-by":"crossref","unstructured":"Zhang X, Li S, Hauer B, Shi N, Kondrak G. Don\u2019t trust ChatGPT when your question is not in English: a study of multilingual abilities and types of LLMs. arXiv:2305.16339. 2023.","DOI":"10.18653\/v1\/2023.emnlp-main.491"},{"key":"ref4","doi-asserted-by":"crossref","first-page":"128089","DOI":"10.1016\/j.neucom.2024.128089","article-title":"BC4LLM: a perspective of trusted artificial intelligence when blockchain meets large language models","volume":"599","author":"Luo","year":"2024","journal-title":"Neurocomputing"},{"key":"ref5","doi-asserted-by":"crossref","unstructured":"Jiao J, Afroogh S, Xu Y, Phillips C. Navigating LLM ethics: advancements, challenges, and future directions. arXiv:2406.18841. 2024.","DOI":"10.1007\/s43681-025-00814-5"},{"key":"ref6","unstructured":"Min B, Ross H, Sulem E, Ben Veyseh AP, Nguyen TH, Sainz O, et al. Recent advances in natural language processing via large pre-trained language models: a survey. arXiv:2111.01243. 2021."},{"key":"ref7","series-title":"Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL)","first-page":"789","article-title":"Pretraining-based natural language generation for text summarization","author":"Zhang","year":"2019"},{"key":"ref8","series-title":"2020 International Conference on Control, Robotics and Intelligent System","first-page":"176","article-title":"Survey on automatic text summarization and transformer models applicability","author":"Wang","year":"2020"},{"key":"ref9","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. arXiv:1706.03762. 2017."},{"key":"ref10","unstructured":"Zaczynska K, Feldhus N, Schwarzenberg R, Gabryszak A, M\u00f6ller S. Evaluating German transformer language models with syntactic agreement tests. arXiv:2007.03765. 2020."},{"key":"ref11","series-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","first-page":"4171","article-title":"BERT: pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2019"},{"key":"ref12","series-title":"Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics","first-page":"4996","article-title":"How multilingual is multilingual BERT?","author":"Pires","year":"2019"},{"key":"ref13","series-title":"Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)","first-page":"5993","article-title":"Abstract text summarization: a low resource challenge","author":"Parida","year":"2019"},{"key":"ref14","doi-asserted-by":"crossref","unstructured":"Scialom T, Dray P-A, Lamprier S, Piwowarski B, Staiano J. MLSUM: the multilingual summarization corpus. arXiv:2004.14900. 2020.","DOI":"10.18653\/v1\/2020.emnlp-main.647"},{"key":"ref15","series-title":"Proceedings of the Twelfth Language Resources and Evaluation Conference","first-page":"6651","article-title":"Summarization corpora of wikipedia articles","author":"Frefel","year":"2020"},{"key":"ref16","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1007\/s10462-016-9475-9","article-title":"Recent automatic text summarization techniques: a survey","volume":"47","author":"Gambhir","year":"2017","journal-title":"Artif Intell Rev"},{"key":"ref17","first-page":"74","author":"Lin","year":"2004","journal-title":"Text summarization branches out"},{"key":"ref18","doi-asserted-by":"crossref","first-page":"101","DOI":"10.1007\/s10579-017-9399-2","article-title":"The challenging task of summary evaluation: an overview","volume":"52","author":"Lloret","year":"2018","journal-title":"Lang Resour Eval"},{"key":"ref19","doi-asserted-by":"crossref","first-page":"43","DOI":"10.1017\/S1351324901002741","article-title":"SUMMAC: a text summarization evaluation","volume":"8","author":"Mani","year":"2002","journal-title":"Nat Lang Eng"},{"key":"ref20","series-title":"Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing","first-page":"419","article-title":"Evaluating information content by factoid analysis: human annotation and stability","author":"Teufel","year":"2004"},{"key":"ref21","series-title":"Proceedings of the Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics: HLT-NAACL","first-page":"145","article-title":"Evaluating content selection in summarization: the pyramid method","author":"Nenkova","year":"2004"},{"key":"ref22","doi-asserted-by":"crossref","first-page":"100058","DOI":"10.1016\/j.health.2022.100058","article-title":"An exploratory study of automatic text summarization in biomedical and healthcare domain","volume":"2","author":"Rohil","year":"2022","journal-title":"Healthc Anal"},{"key":"ref23","series-title":"2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)","first-page":"1155","article-title":"Domain-aware abstractive text summarization for medical documents","author":"Gigioli","year":"2018 Dec 3\u20136"},{"key":"ref24","first-page":"78","article-title":"Towards an efficient approach for automatic medical document summarization","volume":"15","author":"Gayathri","year":"2015","journal-title":"Cybern Inf Technol"},{"key":"ref25","unstructured":"Moradi M, Samwald M. Clustering of deep contextualized representations for summarization of biomedical texts. arXiv:1908.02286. 2019."},{"key":"ref26","first-page":"137","article-title":"German medical natural language processing\u2014a data-centric survey","author":"Zesch","year":"2022","journal-title":"Applicat Med Manufact"},{"key":"ref27","first-page":"24","article-title":"Adapting large language models for automated summarisation of electronic medical records in clinical coding","volume":"318","author":"Bi","year":"2024","journal-title":"Stud Health Technol Inform"},{"key":"ref28","doi-asserted-by":"crossref","first-page":"102904","DOI":"10.1016\/j.artmed.2024.102904","article-title":"Pre-trained language models in medicine: a survey","volume":"154","author":"Luo","year":"2024","journal-title":"Artif Intell Med"},{"key":"ref29","unstructured":"Schubiger R. German summarization with large language models [master\u2019s thesis]. Zurich, Switzerland: ETH Zurich; 2024."},{"key":"ref30","doi-asserted-by":"crossref","unstructured":"Zhang H, Yu PS, Zhang J. A systematic survey of text summarization: from statistical methods to large language models. arXiv:2406.11289. 2024.","DOI":"10.1145\/3731445"},{"key":"ref31","unstructured":"Liddy ED. Natural language processing [Internet]. 2001 [cited 2025 Mar 3]. Available from: https:\/\/surface.syr.edu\/cgi\/viewcontent.cgi?article=1043&context=istpub."},{"key":"ref32","first-page":"33","article-title":"Comparative study of text summarization methods","volume":"102","author":"Munot","year":"2014","journal-title":"Int J Comput Appl"},{"key":"ref33","unstructured":"Liu Y. Fine-tune BERT for extractive summarization. arXiv:1903.10318. 2019."},{"key":"ref34","series-title":"2016 International Conference on Circuit, Power and Computing Technologies (ICCPCT)","first-page":"1","article-title":"A survey on abstractive text summarization","author":"Moratanch","year":"2016 Mar 18\u201319"},{"key":"ref35","unstructured":"Bird S, Klein E, Loper E. Natural language processing with Python: analyzing text with the natural language toolkit. O\u2019Reilly Media [Internet]. 2009 [cited 2025 Mar 3]. Available from: https:\/\/www.nltk.org\/book\/."},{"key":"ref36","unstructured":"Subramanyam Kalyan K, Rajasekharan A, Sangeetha S. AMMUS : a survey of transformer-based pretrained models in natural language processing [Internet]. 2001 [cited 2025 Mar 3]. Available from: https:\/\/mr-nlp.github.io."},{"key":"ref37","series-title":"Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics","first-page":"2978","article-title":"Transformer-XL: attentive language models beyond a fixed-length context","author":"Dai","year":"2019"},{"key":"ref38","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"J Mach Learn Res"},{"key":"ref39","unstructured":"aphasie suisse | Home. Retrieved January 31, 2023. [Internet]. [cited 2025 Mar 3]. Available from: https:\/\/aphasie.org\/."},{"key":"ref40","series-title":"Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics\u2014ACL '04","first-page":"605","article-title":"Automatic evaluation of machine translation quality using longest common subsequence and skip-bigram statistics","author":"Lin","year":"2004 Jul 21\u201326"},{"key":"ref41","series-title":"Proceedings of the Conference on Empirical Methods in Natural Language Processing\u2014EMNLP\u201908","first-page":"186","article-title":"Revisiting readability: a unified framework for predicting text quality","author":"Pitler","year":"2008 Oct 25\u201327"},{"key":"ref42","unstructured":"OpenAI API. Retrieved 2023 Jan 20. [Internet]. [cited 2025 Mar 3]. Available from: https:\/\/openai.com\/api\/."},{"key":"ref43","unstructured":"Hugging Face\u2014The AI community building the future. Retrieved January 20, 2023. [Internet]. [cited 2025 Mar 3]. Available from: https:\/\/huggingface.co\/."},{"key":"ref44","doi-asserted-by":"crossref","unstructured":"Cachola I, Lo K, Cohan A, Weld DS. TLDR: extreme summarization of scientific documents [Internet]. 2020 [cited 2025 Mar 3]. Available from: https:\/\/github.com\/allenai\/scitldr.","DOI":"10.18653\/v1\/2020.findings-emnlp.428"},{"key":"ref45","unstructured":"Schmitter P, Kirecci I, Gatziu Grivas S, Hanne T, Beck C. Transformation Compass f\u00fcr nicht-medizinische Supportprozesse (DE) [Internet]. 2003 [cited 2025 Mar 3]. Available from: https:\/\/digitalcollection.zhaw.ch\/items\/4cf03e1e-6ad0-45b4-8477-86fdf681a416."}],"container-title":["Computers, Materials &amp; Continua"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/cdn.techscience.cn\/files\/cmc\/2025\/TSP_CMC-83-1\/TSP_CMC_61527\/TSP_CMC_61527.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T06:40:11Z","timestamp":1763102411000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.techscience.com\/cmc\/v83n1\/60120"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":45,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025]]},"published-print":{"date-parts":[[2025]]}},"URL":"https:\/\/doi.org\/10.32604\/cmc.2025.061527","relation":{},"ISSN":["1546-2226"],"issn-type":[{"value":"1546-2226","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"2024-11-26","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-03-04","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-03-26","order":2,"name":"published","label":"Published Online","group":{"name":"publication_history","label":"Publication History"}}]}}