{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T22:00:51Z","timestamp":1773525651259,"version":"3.50.1"},"reference-count":35,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,7,2]],"date-time":"2025-07-02T00:00:00Z","timestamp":1751414400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,7,2]],"date-time":"2025-07-02T00:00:00Z","timestamp":1751414400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["82201195"],"award-info":[{"award-number":["82201195"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"National Natural Science Foundation Regional Innovation and Development Joint Fund","award":["U20A20386"],"award-info":[{"award-number":["U20A20386"]}]},{"name":"Key Program of the National Natural Science Foundation of China","award":["82330032"],"award-info":[{"award-number":["82330032"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["npj Digit. Med."],"DOI":"10.1038\/s41746-025-01759-z","type":"journal-article","created":{"date-parts":[[2025,7,2]],"date-time":"2025-07-02T06:19:29Z","timestamp":1751437169000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["Generative artificial intelligence for fundus fluorescein angiography interpretation and human expert evaluation"],"prefix":"10.1038","volume":"8","author":[{"given":"An","family":"Shao","sequence":"first","affiliation":[]},{"given":"Xiaocong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Wenyue","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Yingyu","family":"Li","sequence":"additional","affiliation":[]},{"given":"Hongkang","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Xiangji","family":"Pan","sequence":"additional","affiliation":[]},{"given":"Zexin","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Yufeng","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Tiepei","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Yao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Yih Chung","family":"Tham","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Kai","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Juan","family":"Ye","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,2]]},"reference":[{"key":"1759_CR1","doi-asserted-by":"publisher","first-page":"289","DOI":"10.1111\/ceo.12711","volume":"44","author":"ED Cole","year":"2016","unstructured":"Cole, E. D., Novais, E. A., Louzada, R. N. & Waheed, N. K. Contemporary retinal imaging techniques in diabetic retinopathy: a review. Clin. Exp. Ophthalmol. 44, 289\u2013299 (2016).","journal-title":"Clin. Exp. Ophthalmol."},{"key":"1759_CR2","doi-asserted-by":"publisher","first-page":"1852","DOI":"10.1136\/bjo-2022-321472","volume":"107","author":"Z Gao","year":"2022","unstructured":"Gao, Z. et al. Automatic interpretation and clinical evaluation for fundus fluorescein angiography images of diabetic retinopathy patients by deep learning. Br. J. Ophthalmol. 107, 1852\u20131858 (2022).","journal-title":"Br. J. Ophthalmol."},{"key":"1759_CR3","doi-asserted-by":"publisher","first-page":"588","DOI":"10.1136\/bjophthalmol-2019-314336","volume":"104","author":"S Resnikoff","year":"2020","unstructured":"Resnikoff, S. et al. Estimated number of ophthalmologists worldwide (International Council of Ophthalmology update): will we meet the needs?. Br. J. Ophthalmol. 104, 588\u2013592 (2020).","journal-title":"Br. J. Ophthalmol."},{"key":"1759_CR4","doi-asserted-by":"publisher","first-page":"242","DOI":"10.1001\/jama.2023.25057","volume":"331","author":"MD Howell","year":"2024","unstructured":"Howell, M. D., Corrado, G. S. & Desalvo, K. B. Three epochs of artificial intelligence in health care. JAMA 331, 242\u2013244 (2024).","journal-title":"JAMA"},{"key":"1759_CR5","doi-asserted-by":"publisher","first-page":"1481","DOI":"10.1038\/s41591-024-02959-y","volume":"30","author":"M Christensen","year":"2024","unstructured":"Christensen, M., Vukadinovic, M., Yuan, N. & Ouyang, D. Vision\u2013language foundation model for echocardiogram interpretation. Nat. Med. 30, 1481\u20131488 (2024).","journal-title":"Nat. Med."},{"key":"1759_CR6","doi-asserted-by":"publisher","first-page":"970","DOI":"10.1038\/s41586-024-07894-z","volume":"634","author":"X Wang","year":"2024","unstructured":"Wang, X. et al. A pathology foundation model for cancer diagnosis and prognosis prediction. Nature 634, 970\u2013978 (2024).","journal-title":"Nature"},{"key":"1759_CR7","doi-asserted-by":"publisher","first-page":"466","DOI":"10.1038\/s41586-024-07618-3","volume":"634","author":"MY Lu","year":"2024","unstructured":"Lu, M. Y. et al. A multimodal generative AI copilot for human pathology. Nature 634, 466\u2013473 (2024).","journal-title":"Nature"},{"key":"1759_CR8","doi-asserted-by":"publisher","first-page":"E2229289","DOI":"10.1001\/jamanetworkopen.2022.29289","volume":"5","author":"JS Ahn","year":"2022","unstructured":"Ahn, J. S. et al. Association of artificial intelligence-aided chest radiograph interpretation with reader performance and efficiency. JAMA Netw. Open 5, E2229289 (2022).","journal-title":"JAMA Netw. Open"},{"key":"1759_CR9","doi-asserted-by":"publisher","first-page":"E2255113","DOI":"10.1001\/jamanetworkopen.2022.55113","volume":"6","author":"Y Zhang","year":"2023","unstructured":"Zhang, Y. et al. Comparison of chest radiograph captions based on natural language processing vs completed by radiologists. JAMA Netw. Open 6, E2255113 (2023).","journal-title":"JAMA Netw. Open"},{"key":"1759_CR10","doi-asserted-by":"publisher","first-page":"E2336100","DOI":"10.1001\/jamanetworkopen.2023.36100","volume":"6","author":"J Huang","year":"2023","unstructured":"Huang, J. et al. Generative artificial intelligence for chest radiograph interpretation in the emergency department. JAMA Netw. Open 6, E2336100 (2023).","journal-title":"JAMA Netw. Open"},{"key":"1759_CR11","doi-asserted-by":"crossref","unstructured":"Yu, T. et al. A systematic review of advances in AI-assisted analysis of fundus fluorescein angiography (FFA) images: from detection to report generation. Ophthalmol. Therapy 14, 599\u2013619 (2025).","DOI":"10.1007\/s40123-025-01109-y"},{"key":"1759_CR12","doi-asserted-by":"publisher","DOI":"10.1007\/s10916-023-01987-4","volume":"47","author":"M Koohi-Moghadam","year":"2023","unstructured":"Koohi-Moghadam, M. & Bae, K. T. Generative AI in medical imaging: applications, challenges, and ethics. J. Med. Syst. 47, 94 (2023).","journal-title":"J. Med. Syst."},{"key":"1759_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2022.104220","volume":"135","author":"N Kaur","year":"2022","unstructured":"Kaur, N. & Mittal, A. RadioBERT: a deep learning-based system for medical report generation from chest X-ray images using contextual embeddings. J. Biomed. Inform. 135, 104220 (2022).","journal-title":"J. Biomed. Inform."},{"key":"1759_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.compmedimag.2023.102320","volume":"111","author":"MY Ouis","year":"2024","unstructured":"Ouis, M. Y. & A. Akhloufi, M. Deep learning for report generation on chest X-ray images. Comput. Med. Imaging Graph. 111, 102320 (2024).","journal-title":"Comput. Med. Imaging Graph."},{"key":"1759_CR15","unstructured":"Li, M. et al. FFA-IR: towards an explainable and reliable medical report generation benchmark. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (round 2) (2021)."},{"key":"1759_CR16","doi-asserted-by":"crossref","unstructured":"Li, M. et al. Cross-modal clinical graph transformer for ophthalmic report generation. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 20624-20633 (IEEE, 2022).","DOI":"10.1109\/CVPR52688.2022.02000"},{"key":"1759_CR17","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-024-01101-z","volume":"7","author":"X Chen","year":"2024","unstructured":"Chen, X. et al. FFA-GPT: an automated pipeline for fundus fluorescein angiography interpretation and question-answer. NPJ Digit. Med. 7, 111 (2024).","journal-title":"NPJ Digit. Med."},{"key":"1759_CR18","doi-asserted-by":"publisher","first-page":"1663","DOI":"10.1007\/s00417-021-05503-7","volume":"260","author":"Z Gao","year":"2022","unstructured":"Gao, Z. et al. End-to-end diabetic retinopathy grading based on fundus fluorescein angiography images using deep learning. Graefe\u2019s Arch. Clin. Exp. Ophthalmol. 260, 1663\u20131673 (2022).","journal-title":"Graefe\u2019s Arch. Clin. Exp. Ophthalmol."},{"key":"1759_CR19","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-020-71622-6","volume":"10","author":"K Jin","year":"2020","unstructured":"Jin, K. et al. Automatic detection of non-perfusion areas in diabetic macular edema from fundus fluorescein angiography for decision making using deep learning. Sci. Rep. 10, 15138 (2020).","journal-title":"Sci. Rep."},{"key":"1759_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.aopr.2022.100078","volume":"2","author":"K Jin","year":"2022","unstructured":"Jin, K. & Ye, J. Artificial intelligence and deep learning in ophthalmology: current status and future perspectives. Adv. Ophthalmol. Pract. Res. 2, 100078 (2022).","journal-title":"Adv. Ophthalmol. Pract. Res."},{"key":"1759_CR21","doi-asserted-by":"publisher","first-page":"E231362","DOI":"10.1148\/radiol.231362","volume":"308","author":"MA Fink","year":"2023","unstructured":"Fink, M. A. et al. Potential of ChatGPT and GPT-4 for data mining of free-text CT reports on lung cancer. Radiology 308, E231362 (2023).","journal-title":"Radiology"},{"key":"1759_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2023.102798","volume":"86","author":"S Yang","year":"2023","unstructured":"Yang, S. et al. Radiology report generation with a learned knowledge base and multi-modal alignment. Med. Image Anal. 86, 102798 (2023).","journal-title":"Med. Image Anal."},{"key":"1759_CR23","doi-asserted-by":"publisher","first-page":"E51926","DOI":"10.2196\/51926","volume":"26","author":"X Liu","year":"2024","unstructured":"Liu, X. et al. Uncovering language disparity of ChatGPT on retinal vascular disease classification: cross-sectional study. J. Med. Internet Res. 26, E51926 (2024).","journal-title":"J. Med. Internet Res."},{"key":"1759_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2023.104281","volume":"138","author":"Z Lin","year":"2023","unstructured":"Lin, Z. et al. Contrastive pre-training and linear interaction attention-based transformer for universal medical reports generation. J. Biomed. Inform. 138, 104281 (2023).","journal-title":"J. Biomed. Inform."},{"key":"1759_CR25","doi-asserted-by":"publisher","first-page":"1450","DOI":"10.1136\/bjo-2023-324446","volume":"108","author":"X Chen","year":"2024","unstructured":"Chen, X. et al. ICGA-GPT: Report generation and question answering for indocyanine green angiography images. Br. J. Ophthalmol. 108, 1450\u20131456 (2024).","journal-title":"Br. J. Ophthalmol."},{"key":"1759_CR26","doi-asserted-by":"publisher","first-page":"427","DOI":"10.1016\/j.jmir.2021.04.002","volume":"52","author":"LT Hlabangana","year":"2021","unstructured":"Hlabangana, L. T. et al. Inter-rater reliability in quality assurance (QA) of pediatric chest X-rays. J. Med. Imaging Radiat. Sci. 52, 427\u2013434 (2021).","journal-title":"J. Med. Imaging Radiat. Sci."},{"key":"1759_CR27","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-024-01032-9","volume":"7","author":"Y Wang","year":"2024","unstructured":"Wang, Y. et al. Economic evaluation for medical artificial intelligence: accuracy vs. cost-effectiveness in a diabetic retinopathy screening case. NPJ Digit. Med. 7, 43 (2024).","journal-title":"NPJ Digit. Med."},{"key":"1759_CR28","doi-asserted-by":"publisher","first-page":"783","DOI":"10.1136\/bjophthalmol-2011-301378","volume":"96","author":"S Resnikoff","year":"2012","unstructured":"Resnikoff, S., Felch, W., Gauthier, T. M. & Spivey, B. The number of ophthalmologists in practice and training worldwide: a growing gap despite more than 200 000 practitioners. Br. J. Ophthalmol. 96, 783\u2013787 (2012).","journal-title":"Br. J. Ophthalmol."},{"key":"1759_CR29","doi-asserted-by":"crossref","unstructured":"Chen, Z., Song, Y., Chang, T.-H. & Wan, X. Generating radiology reports via memory-driven transformer. In Proc. 2020 Conference on Empirical Methods in Natural Language Processing 1439\u20131449 (Association for Computational Linguistics, 2020).","DOI":"10.18653\/v1\/2020.emnlp-main.112"},{"key":"1759_CR30","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 770\u2013778 (IEEE, 2016).","DOI":"10.1109\/CVPR.2016.90"},{"key":"1759_CR31","doi-asserted-by":"crossref","unstructured":"Deng, J. et al. ImageNet: a large-scale hierarchical image database. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 248\u2013255 (IEEE, 2009).","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"1759_CR32","doi-asserted-by":"crossref","unstructured":"Qin, H. & Song, Y. Reinforced cross-modal alignment for radiology report generation. In Findings of the Association for Computational Linguistics:ACL 2022 448\u2013458 (Association for Computational Linguistics, 2022).","DOI":"10.18653\/v1\/2022.findings-acl.38"},{"key":"1759_CR33","doi-asserted-by":"crossref","unstructured":"Papineni, K., Roukos, S., Ward, T. & Zhu, W.-J. Bleu: a method for automatic evaluation of machine translation. In Proc. 40th Annual Meeting of the Association for Computational Linguistics 311\u2013318. (Association for Computational Linguistics, 2002).","DOI":"10.3115\/1073083.1073135"},{"key":"1759_CR34","unstructured":"Denkowski, M. & Lavie, A. Meteor 1.3: Automatic Metric for Reliable Optimization and Evaluation of Machine Translation Systems. Proceedings of the Sixth Workshop on Statistical Machine Translation 85\u201391 (Association for Computational Linguistics, 2011)."},{"key":"1759_CR35","unstructured":"Lin, C.-Y. ROUGE: a package for automatic evaluation of summaries. In Text Summarization Branches Out 74\u201381 (Association for Computational Linguistics, 2004)."}],"container-title":["npj Digital Medicine"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01759-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01759-z","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01759-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,2]],"date-time":"2025-07-02T06:19:34Z","timestamp":1751437174000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01759-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,2]]},"references-count":35,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["1759"],"URL":"https:\/\/doi.org\/10.1038\/s41746-025-01759-z","relation":{},"ISSN":["2398-6352"],"issn-type":[{"value":"2398-6352","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,7,2]]},"assertion":[{"value":"4 December 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 May 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 July 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors declare no competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"396"}}