{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,19]],"date-time":"2025-09-19T10:52:29Z","timestamp":1758279149541,"version":"3.40.3"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031189098"},{"type":"electronic","value":"9783031189104"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-18910-4_28","type":"book-chapter","created":{"date-parts":[[2022,10,26]],"date-time":"2022-10-26T23:03:53Z","timestamp":1666825433000},"page":"336-347","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["MultiGAN: Multi-domain Image Translation from\u00a0OCT to\u00a0OCTA"],"prefix":"10.1007","author":[{"given":"Bing","family":"Pan","sequence":"first","affiliation":[]},{"given":"Zexuan","family":"Ji","sequence":"additional","affiliation":[]},{"given":"Qiang","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,27]]},"reference":[{"key":"28_CR1","doi-asserted-by":"publisher","first-page":"3859","DOI":"10.1109\/TMM.2021.3109419","volume":"24","author":"Y Pang","year":"2021","unstructured":"Pang, Y., Lin, J., Qin, T., Chen, Z.: Image-to-image translation: methods and applications. IEEE Trans. Multimed. 24, 3859\u20133881 (2021)","journal-title":"IEEE Trans. Multimed."},{"unstructured":"Kazemi, H., Soleymani, S., Taherkhani, F., Iranmanesh, S., Nasrabadi, N.: Unsupervised image-to-image translation using domain-specific variational information bound. In: Advances in Neural Information Processing Systems, vol. 31 (2018)","key":"28_CR2"},{"key":"28_CR3","doi-asserted-by":"publisher","first-page":"107343","DOI":"10.1016\/j.patcog.2020.107343","volume":"105","author":"R Li","year":"2020","unstructured":"Li, R., Cao, W., Jiao, Q., Wu, S., Wong, H.S.: Simplified unsupervised image translation for semantic segmentation adaptation. Pattern Recogn. 105, 107343 (2020)","journal-title":"Pattern Recogn."},{"key":"28_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"404","DOI":"10.1007\/978-3-030-58529-7_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Cao","year":"2020","unstructured":"Cao, J., Huang, H., Li, Y., He, R., Sun, Z.: Informative sample mining network for multi-domain image-to-image translation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12364, pp. 404\u2013419. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58529-7_24"},{"key":"28_CR5","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"179","DOI":"10.1007\/978-3-030-01219-9_11","volume-title":"Computer Vision \u2013 ECCV 2018","author":"X Huang","year":"2018","unstructured":"Huang, X., Liu, M.-Y., Belongie, S., Kautz, J.: Multimodal unsupervised image-to-image translation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11207, pp. 179\u2013196. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01219-9_11"},{"issue":"4","key":"28_CR6","doi-asserted-by":"publisher","first-page":"e0215076","DOI":"10.1371\/journal.pone.0215076","volume":"14","author":"HL Yang","year":"2019","unstructured":"Yang, H.L., et al.: Weakly supervised lesion localization for age-related macular degeneration detection using optical coherence tomography images. PLoS ONE 14(4), e0215076 (2019)","journal-title":"PLoS ONE"},{"issue":"1","key":"28_CR7","doi-asserted-by":"publisher","first-page":"84","DOI":"10.4103\/2008-322X.180709","volume":"11","author":"K Chalam","year":"2016","unstructured":"Chalam, K., Sambhav, K.: Optical coherence tomography angiography in retinal diseases. J. Ophthalmic Vis. Res. 11(1), 84 (2016)","journal-title":"J. Ophthalmic Vis. Res."},{"key":"28_CR8","doi-asserted-by":"publisher","first-page":"100951","DOI":"10.1016\/j.preteyeres.2021.100951","volume":"84","author":"I La\u0131ns","year":"2021","unstructured":"La\u0131ns, I., et al.: Retinal applications of swept source optical coherence tomography (OCT) and optical coherence tomography angiography (OCTA). Prog. Retin. Eye Res. 84, 100951 (2021)","journal-title":"Prog. Retin. Eye Res."},{"unstructured":"Li, M., et al.: IPN-V2 and OCTA-500: methodology and dataset for retinal image segmentation. arXiv preprint arXiv:2012.07261 (2020)","key":"28_CR9"},{"issue":"1","key":"28_CR10","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1038\/s41598-018-37186-2","volume":"9","author":"CS Lee","year":"2019","unstructured":"Lee, C.S., et al.: Generating retinal flow maps from structural optical coherence tomography with artificial intelligence. Sci. Rep. 9(1), 1\u201311 (2019)","journal-title":"Sci. Rep."},{"key":"28_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"42","DOI":"10.1007\/978-3-030-88013-2_4","volume-title":"Pattern Recognition and Computer Vision","author":"Z Zhang","year":"2021","unstructured":"Zhang, Z., Ji, Z., Chen, Q., Yuan, S., Fan, W.: Texture-guided U-Net for OCT-to-OCTA generation. In: Ma, H., et al. (eds.) PRCV 2021. LNCS, vol. 13022, pp. 42\u201352. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-88013-2_4"},{"issue":"5","key":"28_CR12","doi-asserted-by":"publisher","first-page":"1322","DOI":"10.3390\/jcm9051322","volume":"9","author":"S Kadomoto","year":"2020","unstructured":"Kadomoto, S., Uji, A., Muraoka, Y., Akagi, T., Tsujikawa, A.: Enhanced visualization of retinal microvasculature in optical coherence tomography angiography imaging via deep learning. J. Clin. Med. 9(5), 1322 (2020)","journal-title":"J. Clin. Med."},{"issue":"11","key":"28_CR13","first-page":"1784","volume":"11","author":"XX Li","year":"2018","unstructured":"Li, X.X., et al.: A quantitative comparison of five optical coherence tomography angiography systems in clinical performance. Int. J. Ophthalmol. 11(11), 1784 (2018)","journal-title":"Int. J. Ophthalmol."},{"issue":"10","key":"28_CR14","doi-asserted-by":"publisher","first-page":"5464","DOI":"10.1118\/1.4962470","volume":"43","author":"Q Chen","year":"2016","unstructured":"Chen, Q., Niu, S., Yuan, S., Fan, W., Liu, Q.: High-low reflectivity enhancement based retinal vessel projection for SD-OCT images. Med. Phys. 43(10), 5464\u20135474 (2016)","journal-title":"Med. Phys."},{"doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223\u20132232 (2017)","key":"28_CR15","DOI":"10.1109\/ICCV.2017.244"},{"doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125\u20131134 (2017)","key":"28_CR16","DOI":"10.1109\/CVPR.2017.632"},{"doi-asserted-by":"crossref","unstructured":"Choi, Y., Uh, Y., Yoo, J., Ha, J.W.: StarGAN v2: diverse image synthesis for multiple domains. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8188\u20138197 (2020)","key":"28_CR17","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"28_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"341","DOI":"10.1007\/978-3-030-20873-8_22","volume-title":"Computer Vision \u2013 ACCV 2018","author":"X Yu","year":"2019","unstructured":"Yu, X., Cai, X., Ying, Z., Li, T., Li, G.: SingleGAN: image-to-image translation by a single-generator network using multiple generative adversarial learning. In: Jawahar, C.V., Li, H., Mori, G., Schindler, K. (eds.) ACCV 2018. LNCS, vol. 11365, pp. 341\u2013356. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-20873-8_22"},{"issue":"12","key":"28_CR19","doi-asserted-by":"publisher","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","volume":"39","author":"V Badrinarayanan","year":"2017","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R.: SegNet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481\u20132495 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"28_CR20","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"179","DOI":"10.1007\/978-3-319-46976-8_19","volume-title":"Deep Learning and Data Labeling for Medical Applications","author":"M Drozdzal","year":"2016","unstructured":"Drozdzal, M., Vorontsov, E., Chartrand, G., Kadoury, S., Pal, C.: The importance of skip connections in biomedical image segmentation. In: Carneiro, G., et al. (eds.) LABELS\/DLMIA -2016. LNCS, vol. 10008, pp. 179\u2013187. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46976-8_19"},{"doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. CoRR abs\/1505.04597. arXiv preprint arXiv:1505.04597 (2015)","key":"28_CR21","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"10","key":"28_CR22","doi-asserted-by":"publisher","first-page":"6111","DOI":"10.1007\/s00521-019-04097-w","volume":"32","author":"L Wen","year":"2020","unstructured":"Wen, L., Li, X., Gao, L.: A transfer convolutional neural network for fault diagnosis based on ResNet-50. Neural Comput. Appl. 32(10), 6111\u20136124 (2020). https:\/\/doi.org\/10.1007\/s00521-019-04097-w","journal-title":"Neural Comput. Appl."},{"unstructured":"Park, J., Woo, S., Lee, J.Y., Kweon, I.S.: BAM: bottleneck attention module. arXiv preprint arXiv:1807.06514 (2018)","key":"28_CR23"},{"doi-asserted-by":"crossref","unstructured":"Garg, A., Gowda, D., Kumar, A., Kim, K., Kumar, M., Kim, C.: Improved multi-stage training of online attention-based encoder-decoder models. In: 2019 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), pp. 70\u201377. IEEE (2019)","key":"28_CR24","DOI":"10.1109\/ASRU46091.2019.9003936"},{"issue":"12","key":"28_CR25","doi-asserted-by":"publisher","first-page":"2536","DOI":"10.1109\/TMI.2017.2708987","volume":"36","author":"JM Wolterink","year":"2017","unstructured":"Wolterink, J.M., Leiner, T., Viergever, M.A., I\u0161gum, I.: Generative adversarial networks for noise reduction in low-dose CT. IEEE Trans. Med. Imaging 36(12), 2536\u20132545 (2017)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"3","key":"28_CR26","doi-asserted-by":"publisher","first-page":"8","DOI":"10.4236\/jcc.2019.73002","volume":"7","author":"U Sara","year":"2019","unstructured":"Sara, U., Akter, M., Uddin, M.S.: Image quality assessment through FSIM, SSIM, MSE and PSNR-a comparative study. J. Comput. Commun. 7(3), 8\u201318 (2019)","journal-title":"J. Comput. Commun."},{"doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8789\u20138797 (2018)","key":"28_CR27","DOI":"10.1109\/CVPR.2018.00916"},{"doi-asserted-by":"crossref","unstructured":"Li, X., et al.: Image-to-image translation via hierarchical style disentanglement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8639\u20138648 (2021)","key":"28_CR28","DOI":"10.1109\/CVPR46437.2021.00853"},{"doi-asserted-by":"crossref","unstructured":"Bates, R., Chocholek, M., Fox, C., Howe, J., Jones, N.: SIFID Scottish inshore fisheries integrated data system: WP 3 final report: development of a novel, automated mechanism for the collection of scallop stock data (2020)","key":"28_CR29","DOI":"10.15664\/10023.23449"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-18910-4_28","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,26]],"date-time":"2022-10-26T23:34:13Z","timestamp":1666827253000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-18910-4_28"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031189098","9783031189104"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-18910-4_28","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"27 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shenzhen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/en.prcv.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"microsoft","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"564","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"233","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"41% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.03","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.35","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}