{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T23:56:14Z","timestamp":1743033374354,"version":"3.40.3"},"publisher-location":"Cham","reference-count":27,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031189098"},{"type":"electronic","value":"9783031189104"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-18910-4_3","type":"book-chapter","created":{"date-parts":[[2022,10,26]],"date-time":"2022-10-26T23:03:53Z","timestamp":1666825433000},"page":"28-40","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["FundusGAN: A One-Stage Single Input GAN for\u00a0Fundus Synthesis"],"prefix":"10.1007","author":[{"given":"Chao","family":"Cai","sequence":"first","affiliation":[]},{"given":"Xue","family":"Xia","sequence":"additional","affiliation":[]},{"given":"Yuming","family":"Fang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,27]]},"reference":[{"key":"3_CR1","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1109\/RBME.2010.2084567","volume":"3","author":"MD Abr\u00e0moff","year":"2010","unstructured":"Abr\u00e0moff, M.D., Garvin, M.K., Sonka, M.: Retinal imaging and image analysis. IEEE Rev. Biomed. Eng. 3, 169\u2013208 (2010)","journal-title":"IEEE Rev. Biomed. Eng."},{"key":"3_CR2","doi-asserted-by":"publisher","unstructured":"Fiorini, S., et al.: Automatic generation of synthetic retinal fundus images: vascular network. In: 20th Conference on Medical Image Understanding and Analysis (MIUA), pp. 54\u201360. Springer, Leicestershire (2016). https:\/\/doi.org\/10.1016\/j.procs.2016.07.010","DOI":"10.1016\/j.procs.2016.07.010"},{"key":"3_CR3","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: NIPS, pp. 2678\u20132680. MIT, Montreal (2014)"},{"key":"3_CR4","doi-asserted-by":"crossref","unstructured":"Costa, P., Galdran, A., Meyer, M.I., Abramoff, M.D., Niemeijer, M., Mendon\u00e7a, A., Campilho, A.: Towards adversarial retinal image synthesis. arXiv preprint arXiv:1701.08974 (2017)","DOI":"10.1109\/TMI.2017.2759102"},{"issue":"3","key":"3_CR5","doi-asserted-by":"publisher","first-page":"781","DOI":"10.1109\/TMI.2017.2759102","volume":"37","author":"P Costa","year":"2018","unstructured":"Costa, P., Galdran, A., Meyer, M.I., Niemeijer, M., Abr\u00e0moff, M., Mendon\u00e7a, A.M., Campilho, A.: End-to-end adversarial retinal image synthesis. IEEE Trans. Med. Imaging 37(3), 781\u2013791 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"3_CR6","unstructured":"Kurach, K., Lu\u010di\u0107, M., Zhai, X., Michalski, M., Gelly, S.: A large-scale study on regularization and normalization in GANs. In: ICML (2019)"},{"key":"3_CR7","unstructured":"Roth, K., Lucchi, A., Nowozin, S., Hofmann, T.: Stabilizing training of generative adversarial networks through regularization. In: NIPS, pp. 2015\u20132025. MIT, California (2017)"},{"key":"3_CR8","unstructured":"Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X.: Improved techniques for training GANs. In: NIPS, pp. 2234\u20132242. MIT, Barcelona (2016)"},{"key":"3_CR9","unstructured":"Denton, E., Chintala, S., Szlam, A., Fergus, R.: Deep generative image models using a Laplacian pyramid of adversarial networks. In: NIPS, pp. 1486\u20131494. MIT, Barcelona (2016)"},{"key":"3_CR10","unstructured":"Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: ICML (2017)"},{"key":"3_CR11","unstructured":"Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., Courville, A.: Improved training of Wasserstein GANs. In: NIPS, pp. 5769\u20135779. MIT, California (2017)"},{"key":"3_CR12","unstructured":"Jiang, Y., Chang, S., Wang, Z.: TransGAN: two pure transformers can make one strong GAN, and that can scale up. In: NIPS, pp. 14745\u201314758. Curran Associates Inc, New Orleans (2021)"},{"key":"3_CR13","doi-asserted-by":"crossref","unstructured":"Kamran, S.A., Hossain, K.F., Tavakkoli, A., Zuckerbrod, S.L., Baker, S.A.: VTGAN: semi-supervised retinal image synthesis and disease prediction using vision Transformers. In: ICCV (2021)","DOI":"10.1109\/ICCVW54120.2021.00362"},{"issue":"1","key":"3_CR14","first-page":"1","volume":"10","author":"Z Yu","year":"2019","unstructured":"Yu, Z., Xiang, Q., Meng, J., Kou, C., Ren, Q., Lu, Y.: Retinal image synthesis from multiple-landmarks input with generative adversarial networks. Biomed. Eng. Online 10(1), 1\u201315 (2019)","journal-title":"Biomed. Eng. Online"},{"key":"3_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"235","DOI":"10.1007\/978-3-030-21074-8_19","volume-title":"Computer Vision \u2013 ACCV 2018 Workshops","author":"Y-C Liu","year":"2019","unstructured":"Liu, Y.-C., et al.: Synthesizing new retinal symptom images by multiple generative models. In: Carneiro, G., You, S. (eds.) ACCV 2018. LNCS, vol. 11367, pp. 235\u2013250. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-21074-8_19"},{"key":"3_CR16","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1016\/j.media.2018.07.001","volume":"49","author":"Z He","year":"2018","unstructured":"He, Z., Huiqi, L., Sebastian, M., Li, C.: Synthesizing retinal and neuronal images with generative adversarial nets. Med. Image Anal. 49, 14\u201326 (2018)","journal-title":"Med. Image Anal."},{"key":"3_CR17","doi-asserted-by":"crossref","unstructured":"Sengupta, S., Athwale, A., Gulati, T., Zelek, J., Lakshminarayanan, V.: FunSyn-Net: enhanced residual variational auto-encoder and image-to-image translation network for fundus image synthesis. In: Medical Imaging 2020: Image Processing, vol. 11313, pp. 15\u201310 (2020)","DOI":"10.1117\/12.2549869"},{"issue":"5","key":"3_CR18","first-page":"2567","volume":"445","author":"K Ding","year":"2022","unstructured":"Ding, K., Ma, K., Wang, S., Simoncelli, E.P.: Image quality assessment: unifying structure and texture similarity. IEEE Trans. Pattern Anal. Mach. Intell. 445(5), 2567\u20132581 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"4","key":"3_CR19","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"3_CR20","doi-asserted-by":"publisher","first-page":"511","DOI":"10.1016\/j.ins.2019.06.011","volume":"501","author":"T Li","year":"2019","unstructured":"Li, T., Gao, Y., Wang, K., Guo, S., Liu, H., Kang, H.: Diagnostic assessment of deep learning algorithms for diabetic retinopathy screening. Inf. Sci. 501, 511\u2013522 (2019)","journal-title":"Inf. Sci."},{"key":"3_CR21","doi-asserted-by":"publisher","first-page":"231","DOI":"10.5566\/ias.1155","volume":"33","author":"E Decenci\u00e8re","year":"2014","unstructured":"Decenci\u00e8re, E., et al.: Feedback on a publicly distributed database: the Messidor database. Image Anal. Stereol. 33, 231\u2013234 (2014)","journal-title":"Image Anal. Stereol."},{"key":"3_CR22","doi-asserted-by":"publisher","first-page":"2071","DOI":"10.1109\/JBHI.2020.3028180","volume":"25","author":"K Li","year":"2021","unstructured":"Li, K., Qi, X., Luo, Y., Yao, Z., Zhou, X., Sun, M.: Accurate retinal vessel segmentation in color fundus images via fully attention-based networks. IEEE J. Biomed. Health Inform. 25, 2071\u20132081 (2021)","journal-title":"IEEE J. Biomed. Health Inform."},{"key":"3_CR23","unstructured":"Christian L., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: CVPR, pp. 105\u2013114 (2017)"},{"key":"3_CR24","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-Scale update rule converge to a local Nash equilibrium. In: NIPS, pp. 6629\u20136640. Curran Associates Inc., California (2017)"},{"key":"3_CR25","unstructured":"Radford, A., Metz, L., Chintala, S.: Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 (2015)"},{"key":"3_CR26","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing, improving the image quality of StyleGAN. In: CVPR, pp. 8110\u20138119 (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"3_CR27","unstructured":"Zhang, H., Goodfellow, I., Metaxas, D., Odena, A.: Self-attention generative adversarial networks. In: ICML (2019)"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-18910-4_3","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,26]],"date-time":"2022-10-26T23:31:03Z","timestamp":1666827063000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-18910-4_3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031189098","9783031189104"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-18910-4_3","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"27 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shenzhen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/en.prcv.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"microsoft","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"564","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"233","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"41% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.03","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.35","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}