{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,23]],"date-time":"2025-09-23T00:37:22Z","timestamp":1758587842592,"version":"3.44.0"},"publisher-location":"Cham","reference-count":32,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032060037","type":"print"},{"value":"9783032060044","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T00:00:00Z","timestamp":1758499200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T00:00:00Z","timestamp":1758499200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-06004-4_26","type":"book-chapter","created":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T17:22:15Z","timestamp":1758561735000},"page":"259-268","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Beyond One Size Fits All: Customization of\u00a0Radiology Report Generation Methods"],"prefix":"10.1007","author":[{"given":"Tom","family":"van Sonsbeek","sequence":"first","affiliation":[]},{"given":"Arnaud A. A.","family":"Setio","sequence":"additional","affiliation":[]},{"given":"Jung Oh","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Junwoo","family":"Cho","sequence":"additional","affiliation":[]},{"given":"Junha","family":"Kim","sequence":"additional","affiliation":[]},{"given":"Hyeonsoo","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Gunhee","family":"Nam","sequence":"additional","affiliation":[]},{"given":"Laurent","family":"Dillard","sequence":"additional","affiliation":[]},{"given":"Taesoo","family":"Kim","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,22]]},"reference":[{"key":"26_CR1","doi-asserted-by":"crossref","unstructured":"Alam, H.M.T., Srivastav, D., Kadir, M.A., Sonntag, D.: Towards interpretable radiology report generation via concept bottlenecks using a multi-agentic RAG. arXiv preprint arXiv:2412.16086 (2024)","DOI":"10.1007\/978-3-031-88714-7_18"},{"key":"26_CR2","unstructured":"Bannur, S., et al.: MAIRA-2: grounded radiology report generation. arXiv preprint arXiv:2406.04449 (2024)"},{"key":"26_CR3","doi-asserted-by":"crossref","unstructured":"Bu, S., Li, T., Yang, Y., Dai, Z.: Instance-level expert knowledge and aggregate discriminative attention for radiology report generation. In: CVPR, pp. 14194\u201314204 (2024)","DOI":"10.1109\/CVPR52733.2024.01346"},{"key":"26_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2020.101797","volume":"66","author":"A Bustos","year":"2020","unstructured":"Bustos, A., Pertusa, A., Salinas, J.M., Iglesia-Vaya, M.: Padchest: a large chest X-ray image dataset with multi-label annotated reports. Med. Image Anal. 66, 101797 (2020)","journal-title":"Med. Image Anal."},{"key":"26_CR5","unstructured":"Chambon, P., et al.: Chexpert plus: hundreds of thousands of aligned radiology texts, images and patients. arXiv e-prints pp. arXiv\u20132405 (2024)"},{"key":"26_CR6","doi-asserted-by":"crossref","unstructured":"Chen, Z., Song, Y., Chang, T.H., Wan, X.: Generating radiology reports via memory-driven transformer. EMNLP (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.112"},{"key":"26_CR7","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s12880-021-00671-8","volume":"21","author":"EM Davidson","year":"2021","unstructured":"Davidson, E.M., et al.: The reporting quality of natural language processing studies: systematic review of studies of radiology reports. BMC Med. Imaging 21, 1\u201313 (2021)","journal-title":"BMC Med. Imaging"},{"issue":"2","key":"26_CR8","first-page":"304","volume":"23","author":"D Demner-Fushman","year":"2016","unstructured":"Demner-Fushman, D., et al.: Preparing a collection of radiology examinations for distribution and retrieval. JAMIA 23(2), 304\u2013310 (2016)","journal-title":"JAMIA"},{"key":"26_CR9","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: ICLR (2021)"},{"key":"26_CR10","unstructured":"Dubey, A., et al.: The Llama 3 herd of models. arXiv preprint arXiv:2407.21783 (2024)"},{"issue":"1","key":"26_CR11","first-page":"1","volume":"3","author":"Y Gu","year":"2021","unstructured":"Gu, Y., et al.: Domain-specific language model pretraining for biomedical natural language processing. ACM Health 3(1), 1\u201323 (2021)","journal-title":"ACM Health"},{"key":"26_CR12","doi-asserted-by":"crossref","unstructured":"Hartsock, I., Araujo, C., Folio, L., Rasool, G.: Improving radiology report conciseness and structure via local large language models. arXiv preprint arXiv:2411.05042 (2024)","DOI":"10.1007\/s10278-025-01510-w"},{"key":"26_CR13","unstructured":"Jain, S., et al.: Radgraph: extracting clinical entities and relations from radiology reports. NeurIPS Benchmarks Track (Round 1) (2021)"},{"key":"26_CR14","doi-asserted-by":"crossref","unstructured":"Johnson, A.E., et al.: MIMIC-CXR, a de-identified publicly available database of chest radiographs with free-text reports. Sci. Data 6(1), 317 (2019)","DOI":"10.1038\/s41597-019-0322-0"},{"key":"26_CR15","unstructured":"Lewis, P., et al.: Retrieval-augmented generation for knowledge-intensive NLP tasks. NeurIPS 33, 9459\u20139474 (2020)"},{"key":"26_CR16","doi-asserted-by":"crossref","unstructured":"Li, Y., Wang, Z., Liu, Y., Wang, L., Liu, L., Zhou, L.: Kargen: knowledge-enhanced automated radiology report generation using large language models. In: MICCAI, pp. 382\u2013392. Springer (2024)","DOI":"10.1007\/978-3-031-72086-4_36"},{"key":"26_CR17","doi-asserted-by":"crossref","unstructured":"Liu, Z., Zhu, Z., Zheng, S., Zhao, Y., He, K., Zhao, Y.: From observation to concept: a flexible multi-view paradigm for medical report generation. IEEE Trans. Multimedia (2023)","DOI":"10.1109\/TMM.2023.3342691"},{"key":"26_CR18","unstructured":"Liu, Z., He, K.: A decade\u2019s battle on dataset bias: are we there yet? ICLR (2025)"},{"issue":"3","key":"26_CR19","doi-asserted-by":"publisher","DOI":"10.1148\/radiol.241736","volume":"313","author":"A Meddeb","year":"2024","unstructured":"Meddeb, A., et al.: Large language model ability to translate CT and MRI free-text radiology reports into multiple languages. Radiology 313(3), e241736 (2024)","journal-title":"Radiology"},{"key":"26_CR20","unstructured":"Ranjit, M., Ganapathy, G., Manuel, R., Ganu, T.: Retrieval augmented chest X-ray report generation using OpenAI GPT models. In: Machine Learning for Healthcare Conference, pp. 650\u2013666. PMLR (2023)"},{"key":"26_CR21","unstructured":"Saab, K., et al.: Capabilities of gemini models in medicine. arXiv preprint arXiv:2404.18416 (2024)"},{"key":"26_CR22","doi-asserted-by":"crossref","unstructured":"van Sonsbeek, T., Worring, M.: X-TRA: Improving chest X-ray tasks with cross-modal retrieval augmentation. In: IPMI, pp. 471\u2013482. Springer (2023)","DOI":"10.1007\/978-3-031-34048-2_36"},{"key":"26_CR23","doi-asserted-by":"crossref","unstructured":"Sun, L., Zhao, J., Han, M., Xiong, C.: Fact-aware multimodal retrieval augmentation for accurate medical radiology report generation. NAACL (2025)","DOI":"10.18653\/v1\/2025.naacl-long.28"},{"key":"26_CR24","doi-asserted-by":"crossref","unstructured":"Tanida, T., M\u00fcller, P., Kaissis, G., Rueckert, D.: Interactive and explainable region-guided radiology report generation. In: CVPR, pp. 7433\u20137442 (2023)","DOI":"10.1109\/CVPR52729.2023.00718"},{"key":"26_CR25","unstructured":"Tanno, R., et al.: Collaboration between clinicians and vision\u2013language models in radiology report generation. Nat. Med. 1\u201310 (2024)"},{"key":"26_CR26","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: Activating associative disease-aware vision token memory for LLM-based X-ray report generation. arXiv preprint arXiv:2501.03458 (2025)","DOI":"10.1109\/TMI.2025.3603416"},{"key":"26_CR27","doi-asserted-by":"crossref","unstructured":"Yan, B., et al.: Style-aware radiology report generation with radgraph and few-shot prompting. EMNLP Findings (2023)","DOI":"10.18653\/v1\/2023.findings-emnlp.977"},{"key":"26_CR28","unstructured":"Yang, L., et al.: Advancing multimodal medical capabilities of gemini. arXiv preprint arXiv:2405.03162 (2024)"},{"key":"26_CR29","doi-asserted-by":"crossref","unstructured":"Yu, F., et al.: Evaluating progress in automatic chest X-ray radiology report generation. Patterns 4(9) (2023)","DOI":"10.1016\/j.patter.2023.100802"},{"key":"26_CR30","unstructured":"Zhang, X., et al.: ReXrank: a public leaderboard for AI-powered radiology report generation. arXiv preprint arXiv:2411.15122 (2024)"},{"key":"26_CR31","doi-asserted-by":"crossref","unstructured":"Zhao, W., Wu, C., Zhang, X., Zhang, Y., Wang, Y., Xie, W.: Ratescore: a metric for radiology report generation. EMNLP (2024)","DOI":"10.1101\/2024.06.24.24309405"},{"key":"26_CR32","unstructured":"Zhou, H.Y., Adithan, S., Acosta, J.N., Topol, E.J., Rajpurkar, P.: A generalist learner for multifaceted medical image interpretation. arXiv preprint arXiv:2405.07988 (2024)"}],"container-title":["Lecture Notes in Computer Science","AI for Clinical Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-06004-4_26","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T17:22:25Z","timestamp":1758561745000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-06004-4_26"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,22]]},"ISBN":["9783032060037","9783032060044"],"references-count":32,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-06004-4_26","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,9,22]]},"assertion":[{"value":"22 September 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CMLLMs","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Multimodal Large Language Models in Clinical Practice","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Daejeon","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Korea (Republic of)","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 September 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 September 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cmllms2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/clinicalmllms.github.io\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}