{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T11:57:15Z","timestamp":1775563035851,"version":"3.50.1"},"reference-count":27,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T00:00:00Z","timestamp":1772582400000},"content-version":"vor","delay-in-days":62,"URL":"http:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Procedia Computer Science"],"published-print":{"date-parts":[[2026]]},"DOI":"10.1016\/j.procs.2026.03.076","type":"journal-article","created":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T12:39:40Z","timestamp":1774355980000},"page":"997-1004","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["MedFusion-GPT: A Multimodal Generator Combining GPT-2 and BioGPT for Clinical Report Synthesis"],"prefix":"10.1016","volume":"278","author":[{"given":"Hamida","family":"Abdaoui","sequence":"first","affiliation":[]},{"given":"Sabri","family":"Barbaria","sequence":"additional","affiliation":[]},{"given":"Taha","family":"Ridene","sequence":"additional","affiliation":[]},{"given":"Ridha Ben","family":"Salah","sequence":"additional","affiliation":[]},{"given":"Hanene Boussi","family":"Rahmouni","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.procs.2026.03.076_bib1","unstructured":"Y Chelsea W, Tsao-Lun C, Shankeeth V and al (2025). Artificial Intelligence to Assess Dental Findings from Panoramic Radiographs -- A Multinational Study. arXiv:2502.10277v1."},{"key":"10.1016\/j.procs.2026.03.076_bib2","unstructured":"Wenbo Z, Xin L, Dan Z, Meng J, Linlin F and al (2024). A dual-labeled dataset and fusion model for automatic teeth segmentation, numbering, and state assessment on panoramic radiographs.10.1186\/s12903-024-04984-2."},{"key":"10.1016\/j.procs.2026.03.076_bib3","unstructured":"Pablo de A, Mano M and Romuere R Veloso e S (2024). The Automated Generation of Medical Reports from Polydactyly X-ray Images Using CNNs and Transformers. 0.3390\/app14156566."},{"key":"10.1016\/j.procs.2026.03.076_bib4","unstructured":"Warren B. G, MD and Hiroto H, MD, PhD (2023). Reduction of Errors Resulting from Commonly Missed Chest Radiography Findings. 10.1016\/j.chest.2022.12.003."},{"key":"10.1016\/j.procs.2026.03.076_bib5","unstructured":"Pablo V, Orrana S, Deborah M, Ricardo R, Romuere S (2021). Detecting pulmonary diseases using deep features in X-ray images. 10.1016\/j.patcog.2021.108081."},{"key":"10.1016\/j.procs.2026.03.076_bib6","unstructured":"Mohamed Esmail K \u00b7 Ezz El-Din H,\u00b7Marwa A. S. (2020). Cascaded deeplearning classifiers for computer-aided diagnosis of COVID- 19andpneumoniadiseasesinX-rayscans. 10.1007\/s40747-020-00199-4."},{"key":"10.1016\/j.procs.2026.03.076_bib7","unstructured":"Jarupat J, Rapeeporn C and Choosak N (2025). Advanced AI-assisted panoramic radiograph analysis for periodontal prognostication and alveolar bone loss detection.10.3389\/fdmed.2024.1509361."},{"key":"10.1016\/j.procs.2026.03.076_bib8","unstructured":"Renqian L, Liai S, Yingce X,Tao Q, Sheng Zhang Hoifung P and Tie-Yan L. (2023). BioGPT: Generative Pre-trained Transformer for Biomedical Text Generation and Mining. rXiv:2210.10341v3 [cs.CL]."},{"key":"10.1016\/j.procs.2026.03.076_bib9","unstructured":"Jacob D, Ming-W, Chang K, Lee Kristina T. (2019) BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv:1810.04805v2 [cs.CL]."},{"key":"10.1016\/j.procs.2026.03.076_bib10","unstructured":"Yinhan L, Myle O, NamanG, Jingfei D, Mandar J, Danqi C, Omer L. MikeL, Luke Z, V. (2019) arXiv:1907.11692 [cs.CL]."},{"key":"10.1016\/j.procs.2026.03.076_bib11","unstructured":"Kevin C, Minh-Thang L, Quoc V. L, Christopher D. M (2020). Electra: pre-training text encoders as discriminators rather than generators. arXiv:2003.10555v1 [cs.CL]."},{"key":"10.1016\/j.procs.2026.03.076_bib12","unstructured":"Zhilin Y, Zihang D, Yiming Y, Jaime C, Ruslan S, Quoc V. L. (2020). XLNet: Generalized AutoregressivePre-trainingg for Language Understanding. arXiv:1906.08237v2 [cs.CL]."},{"key":"10.1016\/j.procs.2026.03.076_bib13","series-title":"Improving Language Understanding by GenerativePre-trainingg","author":"Alec","year":"2018"},{"key":"10.1016\/j.procs.2026.03.076_bib14","unstructured":"Alec R, Jeffrey W, Rewon C, David L, Dario A, Ilya S (2019). Language Models are Unsupervised Multitask Learners. OpenAI blog, 1(8):9."},{"key":"10.1016\/j.procs.2026.03.076_bib15","unstructured":"TomB. B, Benjamin M, Nick R, Melanie S and others (2020). Language Models are Few-Shot Learners. Advances in neural information processing systems, 33:18771901."},{"key":"10.1016\/j.procs.2026.03.076_bib16","unstructured":"Yifan P, Shankai Y, Zhiyong L. (2019). Transfer Learning in Biomedical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets. 10.18653\/v1\/W19-5006."},{"key":"10.1016\/j.procs.2026.03.076_bib17","doi-asserted-by":"crossref","unstructured":"Jinhyuk L, Sunkyu K, C, Sungdong K, andJaewoo K, Donghyeon K. (2019). BioBERT: a pre-trained biomedical language representation model for biomedical text mining. doi: 10.1093\/bioinformatics\/btz682.","DOI":"10.1093\/bioinformatics\/btz682"},{"key":"10.1016\/j.procs.2026.03.076_bib18","doi-asserted-by":"crossref","unstructured":"Alistair E.W.J, Tom J.P, Lu S, Li W. L, Mengling F, Mohammad G, Benjamin M, Peter S, LeoAnthony C Roger G.M (2016). MIMIC-III,a freely accessible critical care database. doi: 10.1038\/sdata.2016.35.","DOI":"10.1038\/sdata.2016.35"},{"key":"10.1016\/j.procs.2026.03.076_bib19","doi-asserted-by":"crossref","unstructured":"Iz Beltagy K, Lo Arman C. (2019). SCIBERT: APre-trainedd Language Model for Scientific Text. Proceedings of the 2019 Conference on empirical Methods in Natural Language Processing.","DOI":"10.18653\/v1\/D19-1371"},{"key":"10.1016\/j.procs.2026.03.076_bib20","unstructured":"Y,Robertin, H, Michaellucas N, Xiaodong L, Tristan N, Jianfeng G, and Hoifung P. (2021). Domain-Specific Language ModelPre-trainingg for Biomedical Natural Language Processing. arXiv:2007.15779v6 [cs.CL]."},{"key":"10.1016\/j.procs.2026.03.076_bib21","unstructured":"Milad M, Kathrin B, Florian H, Matthias S. (2021). GPT-3 Models are Poor Few-Shot Learners in the Biomedical Domain. arxiv.org\/abs\/2109.02555."},{"key":"10.1016\/j.procs.2026.03.076_bib22","unstructured":"Bernal G and al. (2022). Thinking about GPT-3 In-Context Learning for Biomedical IE? Think Again. arXiv:2203.08410v3 [cs.CL]."},{"key":"10.1016\/j.procs.2026.03.076_bib23","unstructured":"Tom B. B, Benjamin M, Nick R, Melanie S and others (2020). The language Models are Few-Shot Learners. arXiv:2005.14165."},{"key":"10.1016\/j.procs.2026.03.076_bib24","doi-asserted-by":"crossref","unstructured":"Huan Z, Dong Z, Hao T, Muhammad S, Zhaoquan G. (2023) Medical Specialty Classification Based on Semi adversarial Data Augmentation. https:\/\/doi.org\/10.1155\/2023\/4919371.","DOI":"10.1155\/2023\/4919371"},{"key":"10.1016\/j.procs.2026.03.076_bib25","unstructured":"Karen S, Andrew Z (2015). Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv:1409.1556v6."},{"key":"10.1016\/j.procs.2026.03.076_bib26","doi-asserted-by":"crossref","unstructured":"Srikanth T (2019). Transfer learning using VGG-16 with Deep Convolutional Neural Network for Classifying Images. 10.29322\/IJSRP.9.10. 2019.p9420","DOI":"10.29322\/IJSRP.9.10.2019.p9420"},{"key":"10.1016\/j.procs.2026.03.076_bib27","unstructured":"Gustav M\u00fcller F, Jan Moritz N, Firas K and al. (2023). A multimodal comparison of latent denoising diffusion probabilistic models and generative adversarial networks for medical image synthesis. 10.1038\/s41598-023-39278-0."}],"container-title":["Procedia Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S187705092600668X?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S187705092600668X?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T11:11:15Z","timestamp":1775560275000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S187705092600668X"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"references-count":27,"alternative-id":["S187705092600668X"],"URL":"https:\/\/doi.org\/10.1016\/j.procs.2026.03.076","relation":{},"ISSN":["1877-0509"],"issn-type":[{"value":"1877-0509","type":"print"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"MedFusion-GPT: A Multimodal Generator Combining GPT-2 and BioGPT for Clinical Report Synthesis","name":"articletitle","label":"Article Title"},{"value":"Procedia Computer Science","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.procs.2026.03.076","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 The Author(s). Published by Elsevier B.V.","name":"copyright","label":"Copyright"}]}}