{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T14:24:33Z","timestamp":1769523873809,"version":"3.49.0"},"publisher-location":"Cham","reference-count":47,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197864","type":"print"},{"value":"9783031197871","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19787-1_32","type":"book-chapter","created":{"date-parts":[[2022,10,20]],"date-time":"2022-10-20T22:16:11Z","timestamp":1666304171000},"page":"565-580","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":27,"title":["Custom Structure Preservation in\u00a0Face Aging"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3366-6047","authenticated-orcid":false,"given":"Guillermo","family":"Gomez-Trenado","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6927-8930","authenticated-orcid":false,"given":"St\u00e9phane","family":"Lathuili\u00e8re","sequence":"additional","affiliation":[]},{"given":"Pablo","family":"Mesejo","sequence":"additional","affiliation":[]},{"given":"\u00d3scar","family":"Cord\u00f3n","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,21]]},"reference":[{"key":"32_CR1","doi-asserted-by":"crossref","unstructured":"Ak, K.E., Lim, J.H., Tham, J.Y., Kassim, A.A.: Attribute manipulation generative adversarial networks for fashion images. In: IEEE\/CVF ICCV (2019)","DOI":"10.1109\/ICCV.2019.01064"},{"issue":"4","key":"32_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459805","volume":"40","author":"Y Alaluf","year":"2021","unstructured":"Alaluf, Y., Patashnik, O., Cohen-Or, D.: Only a matter of style: age transformation using a style-based regression model. ACM Trans. Graph. 40(4), 1\u201312 (2021)","journal-title":"ACM Trans. Graph."},{"key":"32_CR3","doi-asserted-by":"crossref","unstructured":"Antipov, G., Baccouche, M., Dugelay, J.L.: Face aging with conditional generative adversarial networks. In: IEEE ICIP (2017)","DOI":"10.1109\/ICIP.2017.8296650"},{"key":"32_CR4","unstructured":"Bi\u0144kowski, M., Sutherland, D.J., Arbel, M., Gretton, A.: Demystifying mmd gans. arXiv preprint arXiv:1801.01401 (2018)"},{"key":"32_CR5","doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: Stargan: unified generative adversarial networks for multi-domain image-to-image translation. In: IEEE\/CVF CVPR (2018)","DOI":"10.1109\/CVPR.2018.00916"},{"key":"32_CR6","doi-asserted-by":"crossref","unstructured":"Fu, H., Gong, M., Wang, C., Batmanghelich, K., Zhang, K., Tao, D.: Geometry-consistent generative adversarial networks for one-sided unsupervised domain mapping. In: IEEE\/CVF CVPR (2019)","DOI":"10.1109\/CVPR.2019.00253"},{"issue":"11","key":"32_CR7","doi-asserted-by":"publisher","first-page":"1955","DOI":"10.1109\/TPAMI.2010.36","volume":"32","author":"Y Fu","year":"2010","unstructured":"Fu, Y., Guo, G., Huang, T.S.: Age synthesis and estimation via faces: a survey. IEEE T-PAMI 32(11), 1955\u20131976 (2010)","journal-title":"IEEE T-PAMI"},{"key":"32_CR8","doi-asserted-by":"crossref","unstructured":"He, Z., Kan, M., Shan, S., Chen, X.: S2gan: share aging factors across ages and share aging trends among individuals. In: IEEE\/CVF ICCV (2019)","DOI":"10.1109\/ICCV.2019.00953"},{"key":"32_CR9","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 6629\u20136640 (2017)"},{"key":"32_CR10","doi-asserted-by":"crossref","unstructured":"Huang, X., Liu, M.Y., Belongie, S., Kautz, J.: Multimodal unsupervised image-to-image translation. In: IEEE\/CVF ECCV (2018)","DOI":"10.1007\/978-3-030-01219-9_11"},{"key":"32_CR11","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: IEEE\/CVF CVPR (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"32_CR12","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: IEEE Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"32_CR13","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of gans for improved quality, stability, and variation. In: ICLR (2017)"},{"key":"32_CR14","unstructured":"Karras, T., Aittala, M., Hellsten, J., Laine, S., Lehtinen, J., Aila, T.: Training generative adversarial networks with limited data. arXiv preprint arXiv:2006.06676 (2020)"},{"key":"32_CR15","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: IEEE\/CVF CVPR (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"32_CR16","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: IEEE\/CVF CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"32_CR17","doi-asserted-by":"crossref","unstructured":"Kemelmacher-Shlizerman, I., Suwajanakorn, S., Seitz, S.M.: Illumination-aware age progression. In: IEEE\/CVF CVPR (2014)","DOI":"10.1109\/CVPR.2014.426"},{"key":"32_CR18","doi-asserted-by":"crossref","unstructured":"Kim, D., Khan, M.A., Choo, J.: Not just compete, but collaborate: local image-to-image translation via cooperative mask prediction. In: IEEE\/CVF CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00644"},{"key":"32_CR19","unstructured":"Lample, G., Zeghidour, N., Usunier, N., Bordes, A., Denoyer, L., et al.: Fader networks: manipulating images by sliding attributes. In: Neurips (2017)"},{"key":"32_CR20","doi-asserted-by":"crossref","unstructured":"Lee, H.Y., Tseng, H.Y., Huang, J.B., Singh, M., Yang, M.H.: Diverse image-to-image translation via disentangled representations. In: IEEE\/CVF ECCV (2018)","DOI":"10.1007\/978-3-030-01246-5_3"},{"key":"32_CR21","unstructured":"Liu, M.Y., Breuel, T., Kautz, J.: Unsupervised image-to-image translation networks. In: Neurips (2017)"},{"key":"32_CR22","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: IEEE\/CVF ICCV (2015)","DOI":"10.1109\/ICCV.2015.425"},{"key":"32_CR23","doi-asserted-by":"crossref","unstructured":"Makhmudkhujaev, F., Hong, S., Park, I.K.: Re-aging gan: toward personalized face age transformation. In: IEEE\/CVF ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00388"},{"key":"32_CR24","unstructured":"Miyato, T., Koyama, M.: cGANs with projection discriminator. arXiv preprint arXiv:1802.05637 (2018)"},{"key":"32_CR25","doi-asserted-by":"crossref","unstructured":"Muhammad, M.B., Yeasin, M.: Eigen-cam: class activation map using principal components. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1\u20137. IEEE (2020)","DOI":"10.1109\/IJCNN48605.2020.9206626"},{"key":"32_CR26","doi-asserted-by":"crossref","unstructured":"Or-El, R., Sengupta, S., Fried, O., Shechtman, E., Kemelmacher-Shlizerman, I.: Lifespan age transformation synthesis. In: IEEE\/CVF ECCV (2020)","DOI":"10.1007\/978-3-030-58539-6_44"},{"key":"32_CR27","doi-asserted-by":"crossref","unstructured":"Pan, H., Han, H., Shan, S., Chen, X.: Mean-variance loss for deep age estimation from a face. In: IEEE\/CVF CVPR (2018)","DOI":"10.1109\/CVPR.2018.00554"},{"key":"32_CR28","first-page":"7198","volume":"33","author":"T Park","year":"2020","unstructured":"Park, T., et al.: Swapping autoencoder for deep image manipulation. Adv. Neural Inf. Process. Syst. 33, 7198\u20137211 (2020)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"32_CR29","doi-asserted-by":"crossref","unstructured":"Pumarola, A., Agudo, A., Martinez, A.M., Sanfeliu, A., Moreno-Noguer, F.: Ganimation: anatomically-aware facial animation from a single image. In: IEEE\/CVF ECCV (2018)","DOI":"10.1007\/978-3-030-01249-6_50"},{"key":"32_CR30","doi-asserted-by":"crossref","unstructured":"Richardson, E., et al.: Encoding in style: a StyleGAN encoder for image-to-image translation. In: IEEE Conference on Computer Vision and Pattern Recognition (2021)","DOI":"10.1109\/CVPR46437.2021.00232"},{"key":"32_CR31","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"issue":"2\u20134","key":"32_CR32","doi-asserted-by":"publisher","first-page":"144","DOI":"10.1007\/s11263-016-0940-3","volume":"126","author":"R Rothe","year":"2018","unstructured":"Rothe, R., Timofte, R., Gool, L.V.: Deep expectation of real and apparent age from a single image without facial landmarks. Int. J. Comput. Vis. 126(2\u20134), 144\u2013157 (2018)","journal-title":"Int. J. Comput. Vis."},{"key":"32_CR33","doi-asserted-by":"crossref","unstructured":"Rothe, R., Timofte, R., Van Gool, L.: Dex: deep expectation of apparent age from a single image. In: IEEE\/CVF ICCV-W (2015)","DOI":"10.1109\/ICCVW.2015.41"},{"key":"32_CR34","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-cam: visual explanations from deep networks via gradient-based localization. In: IEEE\/CVF ICCV (2017)","DOI":"10.1109\/ICCV.2017.74"},{"key":"32_CR35","doi-asserted-by":"crossref","unstructured":"Siarohin, A., Sangineto, E., Lathuiliere, S., Sebe, N.: Deformable gans for pose-based human image generation. In: IEEE\/CVF CVPR (2018)","DOI":"10.1109\/CVPR.2018.00359"},{"key":"32_CR36","unstructured":"Springenberg, J.T., Dosovitskiy, A., Brox, T., Riedmiller, M.A.: Striving for simplicity: the all convolutional net. In: 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7\u20139 May 2015, Workshop Track Proceedings (2015)"},{"key":"32_CR37","unstructured":"Srinivas, S., Fleuret, F.: Full-gradient representation for neural network visualization. Adv. Neural Inf. Process. Syst. 32 (2019)"},{"key":"32_CR38","doi-asserted-by":"crossref","unstructured":"Tang, H., Xu, D., Sebe, N., Yan, Y.: Attention-guided generative adversarial networks for unsupervised image-to-image translation. In: IJCNN (2019)","DOI":"10.1109\/IJCNN.2019.8851881"},{"key":"32_CR39","doi-asserted-by":"crossref","unstructured":"Wang, W., et al.: Recurrent face aging. In: IEEE\/CVF CVPR (2016)","DOI":"10.1109\/CVPR.2016.261"},{"key":"32_CR40","doi-asserted-by":"crossref","unstructured":"Wang, Z., Tang, X., Luo, W., Gao, S.: Face aging with identity-preserved conditional generative adversarial networks. In: IEEE\/CVF CVPR (2018)","DOI":"10.1109\/CVPR.2018.00828"},{"key":"32_CR41","doi-asserted-by":"crossref","unstructured":"Yang, H., Huang, D., Wang, Y., Jain, A.K.: Learning face age progression: a pyramid architecture of gans. In: IEEE\/CVF CVPR (2018)","DOI":"10.1109\/CVPR.2018.00011"},{"key":"32_CR42","doi-asserted-by":"crossref","unstructured":"Yao, X., Newson, A., Gousseau, Y., Hellier, P.: A latent transformer for disentangled face editing in images and videos. In: IEEE\/CVF ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01353"},{"key":"32_CR43","doi-asserted-by":"crossref","unstructured":"Yao, X., Puy, G., Newson, A., Gousseau, Y., Hellier, P.: High resolution face age editing. In: IEEE ICPR (2021)","DOI":"10.1109\/ICPR48806.2021.9412383"},{"key":"32_CR44","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"32_CR45","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Song, Y., Qi, H.: Age progression\/regression by conditional adversarial autoencoder. In: IEEE\/CVF CVPR (2017)","DOI":"10.1109\/CVPR.2017.463"},{"key":"32_CR46","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: IEEE\/CVF ICCV (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"32_CR47","unstructured":"Zhu, J.Y., et al.: Multimodal image-to-image translation by enforcing bi-cycle consistency. In: Neurips (2017)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19787-1_32","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T19:25:15Z","timestamp":1710357915000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19787-1_32"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197864","9783031197871"],"references-count":47,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19787-1_32","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"21 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}