{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,23]],"date-time":"2025-09-23T00:37:13Z","timestamp":1758587833616,"version":"3.44.0"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032060037","type":"print"},{"value":"9783032060044","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T00:00:00Z","timestamp":1758499200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T00:00:00Z","timestamp":1758499200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-06004-4_31","type":"book-chapter","created":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T17:21:41Z","timestamp":1758561701000},"page":"310-319","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DuEL-Med: A Dual-Path Enhanced Language Model for\u00a0Clinically-Aware Radiology Report Generation"],"prefix":"10.1007","author":[{"given":"Jin","family":"Kim","sequence":"first","affiliation":[]},{"given":"Matthew S.","family":"Brown","sequence":"additional","affiliation":[]},{"given":"Dan","family":"Ruan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,22]]},"reference":[{"key":"31_CR1","doi-asserted-by":"crossref","unstructured":"Irvin, J., Rajpurkar, P., Ko, M., Yu, Y., et al.: CheXpert: a large chest radiograph dataset with uncertainty labels and expert comparison. Nat. Mach. Intell. (2019)","DOI":"10.1609\/aaai.v33i01.3301590"},{"key":"31_CR2","unstructured":"Jing, B., Xie, P., Meng, E.: On the automatic generation of medical imaging reports. In: Proceedings of IEEE Conference Computer Vision and Pattern Recognition (CVPR) (2018)"},{"key":"31_CR3","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1007\/s13735-021-00218-1","volume":"11","author":"S Suganyadevi","year":"2022","unstructured":"Suganyadevi, S., Seethalakshmi, V., Balasamy, K.: A review on deep learning in medical image analysis. Int. J. Multimedia Inf. Retrieval 11, 19\u201338 (2022)","journal-title":"Int. J. Multimedia Inf. Retrieval"},{"key":"31_CR4","unstructured":"Chen, H., Zhang, Y., Liao, G., et al.: Generative adversarial networks in medical image analysis: a survey. Med. Image Anal. (2020)"},{"key":"31_CR5","unstructured":"Radford, A., Kim, J., Hallacy, C., Ramesh, A., et al.: Learning transferable visual models from natural language supervision. arXiv:2103.00020 (2021)"},{"key":"31_CR6","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems (NeurIPS) (2017)"},{"key":"31_CR7","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805 (2018)"},{"key":"31_CR8","unstructured":"Xu, K., Ba, J., Kiros, R., Cho, K., et al.: Show, attend and tell: neural image caption generation with visual attention. In: Proceedings of International Conference Machine Learning (ICML) (2015)"},{"key":"31_CR9","doi-asserted-by":"crossref","unstructured":"Wang, X., Wang, F., Li, Y., Ma, Q., et al.: CXPMRG-bench: pre-training and benchmarking for x-ray medical report generation on CheXpert plus. arXiv:2410.00379 (2024)","DOI":"10.1109\/CVPR52734.2025.00483"},{"key":"31_CR10","unstructured":"Hyland, S., Bannur, S., Bouzid, K., Castro, D., et al.: MAIRA-1: a specialised large multimodal model for radiology report generation. arXiv:2311.13668 (2023)"},{"key":"31_CR11","unstructured":"Rajpurkar, P., Irvin, J., Zhu, K., Yang, B., et al.: CheXNeXt: radiologist-level pneumonia detection on chest x-rays with deep learning. JAMA (2018)"},{"key":"31_CR12","unstructured":"Zhang, Y., Liu, X., Chen, S.: Uncertainty-aware deep learning for radiology. IEEE J. Biomed. Health Inf. (2022)"},{"key":"31_CR13","doi-asserted-by":"crossref","unstructured":"Chen, Z., Shen, Y., Song, Y., Wan, X.: Cross-modal memory networks for radiology report generation. arXiv:2204.13258 (2022)","DOI":"10.18653\/v1\/2021.acl-long.459"},{"key":"31_CR14","doi-asserted-by":"crossref","unstructured":"Jin, H., Che, H., Lin, Y., Chen, H.: PromptMRG: diagnosis-driven prompts for medical report generation. In: Proceedings of AAAI Conference Artificial Intelligence, vol. 38, pp. 2607\u20132615 (2024)","DOI":"10.1609\/aaai.v38i3.28038"},{"key":"31_CR15","unstructured":"Pellegrini, C., \u00d6zsoy, E., Busam, B., Wiestler, B., et al.: RaDialog: large vision-language models for x-ray reporting and dialog-driven assistance. In: Medical Imaging with Deep Learning (2025)"},{"key":"31_CR16","unstructured":"Gu, A., Dao, T.: Mamba: linear-time sequence modeling with selective state spaces. arXiv:2312.00752 (2023)"},{"key":"31_CR17","unstructured":"Google DeepMind: Gemini 2.5 Flash-Lite Preview [Model Card]. https:\/\/deepmind.google\/models\/gemini\/flash-lite\/ updated 30 July 2025; Accessed 17 June 2025"},{"key":"31_CR18","unstructured":"Hu, E., Shen, Y., Wallis, P., Allen-Zhu, Z., et al.: LoRA: low-rank adaptation of large language models. arXiv:2106.09685 (2021)"},{"key":"31_CR19","unstructured":"Touvron, H., Martin, L., Stone, K., Albert, P., et al.: Llama 2: open foundation and fine-tuned chat models. arXiv:2307.09288 (2023)"},{"key":"31_CR20","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv:1711.05101 (2017)"},{"key":"31_CR21","doi-asserted-by":"crossref","unstructured":"Demner-Fushman, D., Kohli, M., Rosenman, M., Shooshan, S., et al.: Preparing a collection of radiology examinations for distribution and retrieval. J. Am. Med. Inf. Assoc. (2016)","DOI":"10.1093\/jamia\/ocv080"},{"key":"31_CR22","doi-asserted-by":"publisher","first-page":"100557","DOI":"10.1016\/j.imu.2021.100557","volume":"24","author":"O Alfarghaly","year":"2021","unstructured":"Alfarghaly, O., Khaled, R., Elkorany, A., Helal, M., Fahmy, A.: Automated radiology report generation using conditioned transformers. Inf. Med. Unlocked 24, 100557 (2021)","journal-title":"Inf. Med. Unlocked"},{"key":"31_CR23","unstructured":"Banerjee, S., Lavie, A.: METEOR: an automatic metric for MT evaluation with improved correlation with human judgments. In: Proceedings of ACL Workshop on Intrinsic and Extrinsic Evaluation Measures, pp. 65\u201372 (2005)"},{"key":"31_CR24","doi-asserted-by":"crossref","unstructured":"Ostmeier, S., Xu, J., Chen, Z., Varma, M., et al.: GREEN: generative radiology report evaluation and error notation. arXiv:2405.03595 (2024)","DOI":"10.18653\/v1\/2024.findings-emnlp.21"},{"key":"31_CR25","doi-asserted-by":"crossref","unstructured":"Papineni, K., Roukos, S., Ward, T., Zhu, W.: BLEU: a method for automatic evaluation of machine translation. In: Proceedings of 40th Annual Meeting, Association for Computational Linguistics, pp. 311\u2013318 (2002)","DOI":"10.3115\/1073083.1073135"},{"key":"31_CR26","unstructured":"Lin, C.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74\u201381 (2004)"},{"key":"31_CR27","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Wang, B., Liang, W., Li, Y., et al.: SAM-guided enhanced fine-grained encoding with mixed semantic learning for medical image captioning. In: ICASSP 2024, pp. 1731\u20131735 (2024)","DOI":"10.1109\/ICASSP48485.2024.10446878"},{"key":"31_CR28","unstructured":"Google: MedGemma \u2013 model collection on Hugging Face. https:\/\/huggingface.co\/collections\/google\/medgemma-release-680aade845f90bec6a3f60c4 (2025)"},{"key":"31_CR29","unstructured":"Li, C., Wong, C., Zhang, S., Usuyama, N., et al.: LLaVA-Med: training a large language-and-vision assistant for biomedicine in one day. In: Advances in Neural Information Processing Systems, vol. 36, pp. 28541\u201328564 (2023)"}],"container-title":["Lecture Notes in Computer Science","AI for Clinical Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-06004-4_31","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T17:21:53Z","timestamp":1758561713000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-06004-4_31"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,22]]},"ISBN":["9783032060037","9783032060044"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-06004-4_31","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,9,22]]},"assertion":[{"value":"22 September 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors have no competing interests to declare that are relevant to the content of\u00a0this article.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"CMLLMs","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Multimodal Large Language Models in Clinical Practice","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Daejeon","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Korea (Republic of)","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 September 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 September 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cmllms2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/clinicalmllms.github.io\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}