{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T19:54:42Z","timestamp":1766087682068,"version":"3.40.3"},"publisher-location":"Cham","reference-count":33,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030695408"},{"type":"electronic","value":"9783030695415"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-69541-5_16","type":"book-chapter","created":{"date-parts":[[2021,2,25]],"date-time":"2021-02-25T11:03:47Z","timestamp":1614251027000},"page":"258-273","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["Multiple Exemplars-Based Hallucination for Face Super-Resolution and Editing"],"prefix":"10.1007","author":[{"given":"Kaili","family":"Wang","sequence":"first","affiliation":[]},{"given":"Jose","family":"Oramas","sequence":"additional","affiliation":[]},{"given":"Tinne","family":"Tuytelaars","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,2,26]]},"reference":[{"key":"16_CR1","doi-asserted-by":"crossref","unstructured":"Liu, M.Y., et al.: Few-shot unsupervised image-to-image translation. In: IEEE International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.01065"},{"key":"16_CR2","doi-asserted-by":"crossref","unstructured":"Chen, Y., Tai, Y., Liu, X., Shen, C., Yang, J.: FSRNet: end-to-end learning face super-resolution with facial priors. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00264"},{"key":"16_CR3","doi-asserted-by":"crossref","unstructured":"Bulat, A., Tzimiropoulos, G.: Super-FAN: integrated facial landmark localization and super-resolution of real-world low resolution faces in arbitrary poses with GANs. CVPR abs\/1712.02765 (2018)","DOI":"10.1109\/CVPR.2018.00019"},{"key":"16_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"219","DOI":"10.1007\/978-3-030-01240-3_14","volume-title":"Computer Vision \u2013 ECCV 2018","author":"X Yu","year":"2018","unstructured":"Yu, X., Fernando, B., Ghanem, B., Porikli, F., Hartley, R.: Face super-resolution guided by facial component heatmaps. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11213, pp. 219\u2013235. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01240-3_14"},{"key":"16_CR5","unstructured":"Kim, D., Kim, M., Kwon, G., Kim, D. S.: Progressive face super-resolution via attention to facial landmark. In: Proceedings of the 30th British Machine Vision Conference (BMVC) (2019)"},{"key":"16_CR6","doi-asserted-by":"crossref","unstructured":"Yu, X., Fernando, B., Hartley, R., Porikli, F.: Super-resolving very low-resolution face images with supplementary attributes. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00101"},{"key":"16_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1007\/978-3-030-01258-8_18","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Lu","year":"2018","unstructured":"Lu, Y., Tai, Y.-W., Tang, C.-K.: Attribute-guided face generation using conditional CycleGAN. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11216, pp. 293\u2013308. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01258-8_18"},{"key":"16_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"278","DOI":"10.1007\/978-3-030-01261-8_17","volume-title":"Computer Vision \u2013 ECCV 2018","author":"X Li","year":"2018","unstructured":"Li, X., Liu, M., Ye, Y., Zuo, W., Lin, L., Yang, R.: Learning warped guidance for blind face restoration. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11217, pp. 278\u2013296. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01261-8_17"},{"key":"16_CR9","doi-asserted-by":"crossref","unstructured":"Dogan, B., Gu, S., Timofte, R.: Exemplar guided face image super-resolution without facial landmarks. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (2019)","DOI":"10.1109\/CVPRW.2019.00232"},{"key":"16_CR10","unstructured":"Jingwei, X., Nannan, W., Xinbo, G., Li, J.: Residual attribute attention network for face image super-resolution. In: AAAI (2019)"},{"key":"16_CR11","doi-asserted-by":"crossref","unstructured":"Zhang, K., et al.: Super-identity convolutional neural network for face hallucination (2018)","DOI":"10.1007\/978-3-030-01252-6_12"},{"key":"16_CR12","doi-asserted-by":"crossref","unstructured":"Li, X., Li, W., Ren, D., Zhang, H., Wang, M., Zuo, W.: Enhanced blind face restoration with multi-exemplar images and adaptive spatial feature fusion. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00278"},{"key":"16_CR13","unstructured":"Odena, A., Olah, C., Shlens, J.: Conditional image synthesis with auxiliary classifier GANs. In: Proceedings of the 34th International Conference on Machine Learning, PMLR, pp. 2642\u20132651 (2017)"},{"key":"16_CR14","unstructured":"Mirza, M., Osindero, S.: Conditional generative adversarial nets (2014)"},{"key":"16_CR15","unstructured":"Chen, X., Duan, Y., Houthooft, R., Schulman, J., Sutskever, I., Abbeel, P.: InfoGAN: interpretable representation learning by information maximizing generative adversarial nets. In: NIPS (2016)"},{"key":"16_CR16","doi-asserted-by":"crossref","unstructured":"Bao, J., Chen, D., Wen, F., Li, H., Hua, G.: CVAE-GAN: fine-grained image generation through asymmetric training. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.299"},{"key":"16_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"776","DOI":"10.1007\/978-3-319-46493-0_47","volume-title":"Computer Vision \u2013 ECCV 2016","author":"X Yan","year":"2016","unstructured":"Yan, X., Yang, J., Sohn, K., Lee, H.: Attribute2Image: conditional image generation from visual attributes. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9908, pp. 776\u2013791. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46493-0_47"},{"key":"16_CR18","doi-asserted-by":"crossref","unstructured":"Esser, P., Sutter, E., Ommer, B.: A variational U-net for conditional appearance and shape generation (2018)","DOI":"10.1109\/CVPR.2018.00923"},{"key":"16_CR19","doi-asserted-by":"crossref","unstructured":"Lin, J., Xia, Y., Qin, T., Chen, Z., Liu, T.Y.: Conditional image-to-image translation. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00579"},{"key":"16_CR20","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"179","DOI":"10.1007\/978-3-030-01219-9_11","volume-title":"Computer Vision \u2013 ECCV 2018","author":"X Huang","year":"2018","unstructured":"Huang, X., Liu, M.-Y., Belongie, S., Kautz, J.: Multimodal unsupervised image-to-image translation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11207, pp. 179\u2013196. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01219-9_11"},{"key":"16_CR21","doi-asserted-by":"crossref","unstructured":"Liu, Y., Junjie, Y., Ouyang, W.: Quality aware network for set to set recognition. In: IEEE Conference on Computer Vision and Pattern Recognition (2017)","DOI":"10.1109\/CVPR.2017.499"},{"key":"16_CR22","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. CVPR abs\/1812.04948 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"16_CR23","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"16_CR24","doi-asserted-by":"crossref","unstructured":"Albanie, S., Nagrani, A., Vedaldi, A., Zisserman, A.: Emotion recognition in speech using cross-modal transfer in the wild. In: ACM Multimedia (2018)","DOI":"10.1145\/3240508.3240578"},{"key":"16_CR25","unstructured":"Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., Courville, A.C.: Improved training of Wasserstein GANs. In: NIPS (2017)"},{"key":"16_CR26","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: Proceedings of International Conference on Computer Vision (ICCV) (2015)","DOI":"10.1109\/ICCV.2015.425"},{"key":"16_CR27","unstructured":"Yi, D., Lei, Z., Liao, S., Li, S.Z.: Learning face representation from scratch (2014)"},{"key":"16_CR28","unstructured":"Paszke, A., et al.: Automatic differentiation in PyTorch (2017)"},{"key":"16_CR29","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization (2015)"},{"key":"16_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"694","DOI":"10.1007\/978-3-319-46475-6_43","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Johnson","year":"2016","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694\u2013711. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43"},{"key":"16_CR31","doi-asserted-by":"crossref","unstructured":"Hanhart, P., Korshunov, P., Ebrahimi, T.: Benchmarking of quality metrics on ultra-high definition video sequences. In: 2013 18th International Conference on Digital Signal Processing (DSP), pp. 1\u20138 (2013)","DOI":"10.1109\/ICDSP.2013.6622760"},{"key":"16_CR32","unstructured":"Dosovitskiy, A., Brox, T.: Generating images with perceptual similarity metrics based on deep networks (2016)"},{"key":"16_CR33","doi-asserted-by":"crossref","unstructured":"Ledig, C., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 105\u2013114 (2017)","DOI":"10.1109\/CVPR.2017.19"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ACCV 2020"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-69541-5_16","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,2,25]],"date-time":"2021-02-25T11:31:58Z","timestamp":1614252718000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-69541-5_16"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030695408","9783030695415"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-69541-5_16","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"26 February 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ACCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Asian Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kyoto","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 November 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 December 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"accv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/accv2020.kyoto\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"768","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"254","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"33% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}