{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T06:10:05Z","timestamp":1752473405473,"version":"3.41.2"},"publisher-location":"Cham","reference-count":33,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031986932","type":"print"},{"value":"9783031986949","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,7,15]],"date-time":"2025-07-15T00:00:00Z","timestamp":1752537600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,7,15]],"date-time":"2025-07-15T00:00:00Z","timestamp":1752537600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-031-98694-9_21","type":"book-chapter","created":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T05:51:19Z","timestamp":1752472279000},"page":"293-305","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Exploring the Effectiveness of Deep Features from Domain-Specific Foundation Models in Retinal Image Synthesis"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0000-9050-383X","authenticated-orcid":false,"given":"Zuzanna","family":"Sk\u00f3rniewska","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8432-2511","authenticated-orcid":false,"given":"Bart\u0142omiej W.","family":"Papie\u017c","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,15]]},"reference":[{"issue":"3","key":"21_CR1","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1038\/s41551-018-0195-0","volume":"2","author":"R Poplin","year":"2018","unstructured":"Poplin, R., et al.: Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nat. Biomed. Eng. 2(3), 158\u2013164 (2018). https:\/\/doi.org\/10.1038\/s41551-018-0195-0","journal-title":"Nat. Biomed. Eng."},{"issue":"6","key":"21_CR2","doi-asserted-by":"publisher","first-page":"498","DOI":"10.1038\/s41551-020-00626-4","volume":"5","author":"CY Cheung","year":"2021","unstructured":"Cheung, C.Y., et al.: A deep-learning system for the assessment of cardiovascular disease risk via the measurement of retinal-vessel calibre. Nat Biomed Eng. 5(6), 498\u2013508 (2021). https:\/\/doi.org\/10.1038\/s41551-020-00626-4","journal-title":"Nat Biomed Eng."},{"issue":"7981","key":"21_CR3","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1038\/s41586-023-06555-x","volume":"622","author":"Y Zhou","year":"2023","unstructured":"Zhou, Y., et al.: A foundation model for generalizable disease detection from retinal images. Nature 622(7981), 156\u2013163 (2023). https:\/\/doi.org\/10.1038\/s41586-023-06555-x","journal-title":"Nature"},{"issue":"11","key":"21_CR4","doi-asserted-by":"publisher","first-page":"e806","DOI":"10.1016\/S2589-7500(22)00169-8","volume":"4","author":"CY Cheung","year":"2022","unstructured":"Cheung, C.Y., et al.: A deep learning model for detection of Alzheimer\u2019s disease based on retinal photographs: a retrospective, multicentre case-control study. Lancet. Digit. Health 4(11), e806\u2013e815 (2022). https:\/\/doi.org\/10.1016\/S2589-7500(22)00169-8","journal-title":"Lancet. Digit. Health"},{"issue":"1","key":"21_CR5","doi-asserted-by":"publisher","first-page":"4061313","DOI":"10.1155\/2019\/4061313","volume":"2019","author":"G An","year":"2019","unstructured":"An, G., et al.: Glaucoma diagnosis with machine learning based on optical coherence tomography and color fundus images. J. Healthc. Eng. 2019(1), 4061313 (2019). https:\/\/doi.org\/10.1155\/2019\/4061313","journal-title":"J. Healthc. Eng."},{"issue":"7","key":"21_CR6","doi-asserted-by":"publisher","first-page":"12","DOI":"10.1167\/tvst.11.7.12","volume":"11","author":"Y Zhou","year":"2022","unstructured":"Zhou, Y., et al.: AutoMorph: automated retinal vascular morphology quantification via a deep learning pipeline. Transl. Vision Sci. Technol. 11(7), 12 (2022). https:\/\/doi.org\/10.1167\/tvst.11.7.12","journal-title":"Transl. Vision Sci. Technol."},{"issue":"4","key":"21_CR7","doi-asserted-by":"publisher","first-page":"501","DOI":"10.1109\/TMI.2004.825627","volume":"23","author":"J Staal","year":"2004","unstructured":"Staal, J., Abr\u00e0moff, M.D., Niemeijer, M., Viergever, M.A., Van Ginneken, B.: Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 23(4), 501\u2013509 (2004). https:\/\/doi.org\/10.1109\/TMI.2004.825627","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"3","key":"21_CR8","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1109\/42.845178","volume":"19","author":"A Hoover","year":"2000","unstructured":"Hoover, A.: Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response. IEEE Trans. Med. Imaging 19(3), 203\u2013210 (2000). https:\/\/doi.org\/10.1109\/42.845178","journal-title":"IEEE Trans. Med. Imaging"},{"key":"21_CR9","doi-asserted-by":"crossref","unstructured":"Alloula, A., Mustafa, R., McGowan, D.R., Papie\u017c, B.W.: On Biases in a UK Biobank-based Retinal Image Classification Model (2024). https:\/\/arxiv.org\/abs\/2408.02676","DOI":"10.1007\/978-3-031-72787-0_14"},{"key":"21_CR10","doi-asserted-by":"crossref","unstructured":"Esser, P., Rombach, R., Ommer, B.: Taming Transformers for High-Resolution Image Synthesis (2021). https:\/\/arxiv.org\/abs\/2012.09841","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"21_CR11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042","author":"R Rombach","year":"2022","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. Proc. IEEE Comput. Soci. Conf. Comput. Vision Pattern Recognit. (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.01042","journal-title":"Proc. IEEE Comput. Soci. Conf. Comput. Vision Pattern Recognit."},{"key":"21_CR12","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Adv. Neural Inf. Proc. Syst. 33, 6840\u20136851 (2020)","journal-title":"Adv. Neural Inf. Proc. Syst."},{"key":"21_CR13","unstructured":"Song, J., Meng, C., Ermon, S.: Denoising diffusion implicit models. In: ICLR 2021 - 9th International Conference on Learning Representations (2021)"},{"key":"21_CR14","doi-asserted-by":"publisher","unstructured":"Tudosiu, P.-D., et al.: Morphology-preserving Autoregressive 3D Generative Modelling of the Brain (2022). https:\/\/doi.org\/10.48550\/ARXIV.2209.03177","DOI":"10.48550\/ARXIV.2209.03177"},{"key":"21_CR15","doi-asserted-by":"publisher","unstructured":"Zhu, X., Zhang, W., Li, Y., O\u2019Donnell, L.J., Zhang, F.: When Diffusion MRI Meets Diffusion Model: A Novel Deep Generative Model for Diffusion MRI Generation (2024). https:\/\/doi.org\/10.48550\/ARXIV.2408.12897","DOI":"10.48550\/ARXIV.2408.12897"},{"key":"21_CR16","doi-asserted-by":"crossref","unstructured":"Litrico, M., Guarnera, F., Giuffrida, M.V., Rav\u00ec, D., Battiato, S.: TADM: Temporally-aware diffusion model for neurodegenerative progression on brain MRI. In: proceedings of Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2024. Springer, Switzerland (2024)","DOI":"10.1007\/978-3-031-72069-7_42"},{"key":"21_CR17","doi-asserted-by":"crossref","unstructured":"Konz, N., Chen, Y., Dong, H., Mazurowski, M.A.: Anatomically-controllable medical image generation with segmentation-guided diffusion models. In: proceedings of Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2024. Springer Nature, Switzerland (2024)","DOI":"10.1007\/978-3-031-72104-5_9"},{"key":"21_CR18","doi-asserted-by":"crossref","unstructured":"Doerrich, S., Di Salvo, F., Ledig, C.: Self-supervised vision transformer are scalable generative models for domain generalization. In: Proceedings of Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2024. Springer Nature, Switzerland (2024)","DOI":"10.1007\/978-3-031-72117-5_60"},{"key":"21_CR19","doi-asserted-by":"crossref","unstructured":"Sturm, M., Cerrone, L., Hamprecht, F.A.: SynCellFactory: generative data augmentation for cell tracking. In: Proceedings of Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2024. Springer Nature, Switzerland (2024)","DOI":"10.1007\/978-3-031-72390-2_29"},{"key":"21_CR20","doi-asserted-by":"publisher","unstructured":"Zhang, H., et al.: LeFusion: Controllable Pathology Synthesis via Lesion-Focused Diffusion Models (2024). https:\/\/doi.org\/10.48550\/ARXIV.2403.14066","DOI":"10.48550\/ARXIV.2403.14066"},{"key":"21_CR21","doi-asserted-by":"publisher","unstructured":"Bradbury, R., Vallis, K.A., Papiez, B.W.: Paired diffusion: generation of related, synthetic PET-CT-segmentation scans using linked denoising diffusion probabilistic models. In: Proceedings - International Symposium on Biomedical Imaging (2024). https:\/\/doi.org\/10.1109\/ISBI56570.2024.10635593","DOI":"10.1109\/ISBI56570.2024.10635593"},{"key":"21_CR22","doi-asserted-by":"publisher","DOI":"10.1109\/JBHI.2025.3526806","author":"L Zhang","year":"2025","unstructured":"Zhang, L., Wu, F., Bronik, K., Papiez, B.W.: DiffuSeg: domain-driven diffusion for medical image segmentation. IEEE J. Biomed. Health Inform. (2025). https:\/\/doi.org\/10.1109\/JBHI.2025.3526806","journal-title":"IEEE J. Biomed. Health Inform."},{"key":"21_CR23","unstructured":"Hou, Q., et al.: FundusGAN: a Hierarchical Feature-Aware Generative Framework for High-Fidelity Fundus Image Generation (2025). Accessed 05 Apr 2025. https:\/\/arxiv.org\/abs\/2503.17831v1"},{"key":"21_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/J.BSPC.2023.105289","volume":"86","author":"S Ahn","year":"2023","unstructured":"Ahn, S., Song, S.J., Shin, J.: FundusGAN: fundus image synthesis based on semi-supervised learning. Biomed. Signal Process. Control 86, 105289 (2023). https:\/\/doi.org\/10.1016\/J.BSPC.2023.105289","journal-title":"Biomed. Signal Process. Control"},{"key":"21_CR25","doi-asserted-by":"publisher","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00068","DOI":"10.1109\/CVPR.2018.00068"},{"key":"21_CR26","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2015). https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"21_CR27","unstructured":"van den Oord, A., Vinyals, O., Kavukcuoglu, K.: Neural Discrete Representation Learning (2018). https:\/\/arxiv.org\/abs\/1711.00937"},{"key":"21_CR28","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale (2021). https:\/\/arxiv.org\/abs\/2010.11929"},{"issue":"2","key":"21_CR29","doi-asserted-by":"publisher","first-page":"167","DOI":"10.1002\/cyto.a.20022","volume":"58","author":"E Meijering","year":"2004","unstructured":"Meijering, E., Jacob, M., Sarria, J.C.F., Steiner, P., Hirling, H., Unser, M.: Design and validation of a tool for neurite tracing and analysis in fluorescence microscopy images. Cytometry A 58(2), 167\u2013176 (2004). https:\/\/doi.org\/10.1002\/cyto.a.20022","journal-title":"Cytometry A"},{"key":"21_CR30","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: 3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings (2014). Accessed 06 Apr. 2025. https:\/\/arxiv.org\/abs\/1409.1556v6"},{"key":"21_CR31","doi-asserted-by":"publisher","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.308","DOI":"10.1109\/CVPR.2016.308"},{"key":"21_CR32","unstructured":"Jordon, J., et al.: Synthetic Data -- what, why and how?\u2019 (2022)"},{"key":"21_CR33","doi-asserted-by":"publisher","unstructured":"Dong, X., et al.: PeCo: Perceptual codebook for BERT pre-training of vision transformers. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 1, pp. 552\u2013560 (2023). https:\/\/doi.org\/10.1609\/AAAI.V37I1.25130","DOI":"10.1609\/AAAI.V37I1.25130"}],"container-title":["Lecture Notes in Computer Science","Medical Image Understanding and Analysis"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-98694-9_21","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T05:51:24Z","timestamp":1752472284000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-98694-9_21"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,15]]},"ISBN":["9783031986932","9783031986949"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-98694-9_21","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,7,15]]},"assertion":[{"value":"15 July 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MIUA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Annual Conference on Medical Image Understanding and Analysis","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Leeds","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 July 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 July 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miua2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/conferences.leeds.ac.uk\/miua\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}