{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,27]],"date-time":"2026-02-27T11:19:05Z","timestamp":1772191145760,"version":"3.50.1"},"publisher-location":"Cham","reference-count":44,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197864","type":"print"},{"value":"9783031197871","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19787-1_42","type":"book-chapter","created":{"date-parts":[[2022,10,20]],"date-time":"2022-10-20T22:16:11Z","timestamp":1666304171000},"page":"737-754","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":34,"title":["EleGANt: Exquisite and\u00a0Locally Editable GAN for\u00a0Makeup Transfer"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9997-4929","authenticated-orcid":false,"given":"Chenyu","family":"Yang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6256-7908","authenticated-orcid":false,"given":"Wanrong","family":"He","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6874-7026","authenticated-orcid":false,"given":"Yingqing","family":"Xu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0348-2546","authenticated-orcid":false,"given":"Yang","family":"Gao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,21]]},"reference":[{"key":"42_CR1","doi-asserted-by":"crossref","unstructured":"An, J., Xiong, H., Huan, J., Luo, J.: Ultrafast photorealistic style transfer via neural architecture search. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 10443\u201310450 (2020)","DOI":"10.1609\/aaai.v34i07.6614"},{"key":"42_CR2","doi-asserted-by":"crossref","unstructured":"Cai, M., Zhang, H., Huang, H., Geng, Q., Li, Y., Huang, G.: Frequency domain image translation: more photo-realistic, better identity-preserving. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 13930\u201313940 (2021)","DOI":"10.1109\/ICCV48922.2021.01367"},{"key":"42_CR3","doi-asserted-by":"crossref","unstructured":"Chang, H., Lu, J., Yu, F., Finkelstein, A.: PairedCycleGAN: asymmetric style transfer for applying and removing makeup. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 40\u201348 (2018)","DOI":"10.1109\/CVPR.2018.00012"},{"key":"42_CR4","unstructured":"Chen, C.F., Panda, R., Fan, Q.: RegionViT: regional-to-local attention for vision transformers. In: Proceedings of the International Conference on Learning Representations (ICLR) (2022)"},{"key":"42_CR5","doi-asserted-by":"crossref","unstructured":"Chen, H.J., Hui, K.M., Wang, S.Y., Tsao, L.W., Shuai, H.H., Cheng, W.H.: BeautyGlow: on-demand makeup transfer framework with reversible generative network. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10042\u201310050 (2019)","DOI":"10.1109\/CVPR.2019.01028"},{"key":"42_CR6","doi-asserted-by":"crossref","unstructured":"Choi, Y., Uh, Y., Yoo, J., Ha, J.W.: StarGAN v2: diverse image synthesis for multiple domains. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8188\u20138197 (2020)","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"42_CR7","unstructured":"Chu, X., et al.: Twins: revisiting the design of spatial attention in vision transformers. In: Proceedings of the International Conference on Neural Information Processing Systems (NIPS), pp. 9355\u20139366 (2021)"},{"key":"42_CR8","doi-asserted-by":"crossref","unstructured":"Deng, H., Han, C., Cai, H., Han, G., He, S.: Spatially-invariant style-codes controlled makeup transfer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6549\u20136557 (2021)","DOI":"10.1109\/CVPR46437.2021.00648"},{"key":"42_CR9","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: Proceedings of the International Conference on Learning Representations (ICLR) (2021)"},{"key":"42_CR10","doi-asserted-by":"crossref","unstructured":"Gatys, L.A., Ecker, A.S., Bethge, M.: Image style transfer using convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2414\u20132423 (2016)","DOI":"10.1109\/CVPR.2016.265"},{"key":"42_CR11","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: Proceedings of the International Conference on Neural Information Processing Systems (NIPS) (2014)"},{"key":"42_CR12","doi-asserted-by":"crossref","unstructured":"Gu, Q., Wang, G., Chiu, M.T., Tai, Y.W., Tang, C.K.: LADN: local adversarial disentangling network for facial makeup and de-makeup. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 10481\u201310490 (2019)","DOI":"10.1109\/ICCV.2019.01058"},{"key":"42_CR13","doi-asserted-by":"crossref","unstructured":"Guo, D., Sim, T.: Digital face makeup by example. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 73\u201379 (2009)","DOI":"10.1109\/CVPR.2009.5206833"},{"key":"42_CR14","doi-asserted-by":"crossref","unstructured":"Heo, B., Yun, S., Han, D., Chun, S., Choe, J., Oh, S.J.: Rethinking spatial dimensions of vision transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 11936\u201311945 (2021)","DOI":"10.1109\/ICCV48922.2021.01172"},{"key":"42_CR15","unstructured":"Ho, J., Kalchbrenner, N., Weissenborn, D., Salimans, T.: Axial attention in multidimensional transformers. arXiv preprint arXiv:1912.12180 (2019)"},{"key":"42_CR16","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 1501\u20131510 (2017)","DOI":"10.1109\/ICCV.2017.167"},{"key":"42_CR17","unstructured":"Jaderberg, M., Simonyan, K., Zisserman, A., Kavukcuoglu, K.: Spatial transformer networks. In: Proceedings of the International Conference on Neural Information Processing Systems (NIPS), pp. 2017\u20132025 (2015)"},{"key":"42_CR18","doi-asserted-by":"crossref","unstructured":"Jiang, W., et al.: PSGAN: pose and expression robust spatial-aware GAN for customizable makeup transfer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5194\u20135202 (2020)","DOI":"10.1109\/CVPR42600.2020.00524"},{"key":"42_CR19","doi-asserted-by":"crossref","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 694\u2013711 (2016)","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"42_CR20","doi-asserted-by":"crossref","unstructured":"Kim, S.S., Kolkin, N., Salavon, J., Shakhnarovich, G.: Deformable style transfer. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 246\u2013261 (2020)","DOI":"10.1007\/978-3-030-58574-7_15"},{"key":"42_CR21","doi-asserted-by":"crossref","unstructured":"Li, C., Zhou, K., Lin, S.: Simulating makeup through physics-based manipulation of intrinsic image layers. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4621\u20134629 (2015)","DOI":"10.1109\/CVPR.2015.7299093"},{"key":"42_CR22","doi-asserted-by":"crossref","unstructured":"Li, T., et al.: BeautyGAN: instance-level facial makeup transfer with deep generative adversarial network. In: Proceedings of the 26th ACM International Conference on Multimedia, pp. 645\u2013653 (2018)","DOI":"10.1145\/3240508.3240618"},{"issue":"4","key":"42_CR23","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073683","volume":"36","author":"J Liao","year":"2017","unstructured":"Liao, J., Yao, Y., Yuan, L., Hua, G., Kang, S.B.: Visual attribute transfer through deep image analogy. ACM Trans. Graph. 36(4), 1\u201315 (2017)","journal-title":"ACM Trans. Graph."},{"key":"42_CR24","doi-asserted-by":"crossref","unstructured":"Liu, L., Xing, J., Liu, S., Xu, H., Zhou, X., Yan, S.: Wow! You are so beautiful today! ACM Trans. Multim. Comput. Commun. Appl. (TOMM) 11(1s), 1\u201322 (2014)","DOI":"10.1145\/2659234"},{"key":"42_CR25","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"42_CR26","doi-asserted-by":"crossref","unstructured":"Luan, F., Paris, S., Shechtman, E., Bala, K.: Deep photo style transfer. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 4990\u20134998 (2017)","DOI":"10.1109\/CVPR.2017.740"},{"key":"42_CR27","unstructured":"Parmar, N., et al.: Image transformer. In: Proceedings of the International Conference on Machine Learning (ICML), pp. 4055\u20134064 (2018)"},{"key":"42_CR28","doi-asserted-by":"crossref","unstructured":"Rakhimov, R., Volkhonskiy, D., Artemov, A., Zorin, D., Burnaev, E.: Latent video transformer. In: Proceedings of the 16th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP), pp. 101\u2013112 (2021)","DOI":"10.5220\/0010241801010112"},{"issue":"9","key":"42_CR29","doi-asserted-by":"publisher","first-page":"2035","DOI":"10.1109\/TPAMI.2018.2848939","volume":"41","author":"B Shi","year":"2018","unstructured":"Shi, B., Yang, M., Wang, X., Lyu, P., Yao, C., Bai, X.: Aster: an attentional scene text recognizer with flexible rectification. IEEE Trans. Pattern Anal. Mach. Intell. 41(9), 2035\u20132048 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"42_CR30","doi-asserted-by":"crossref","unstructured":"Tong, W.S., Tang, C.K., Brown, M.S., Xu, Y.Q.: Example-based cosmetic transfer. In: Proceedings of the 15th Pacific Conference on Computer Graphics and Applications (PG), pp. 211\u2013218 (2007)","DOI":"10.1109\/PG.2007.31"},{"key":"42_CR31","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Proceedings of the International Conference on Neural Information Processing Systems (NIPS), pp. 6000\u20136010 (2017)"},{"key":"42_CR32","doi-asserted-by":"crossref","unstructured":"Wan, Z., Chen, H., An, J., Jiang, W., Yao, C., Luo, J.: Facial attribute transformers for precise and robust makeup transfer. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 1717\u20131726 (2022)","DOI":"10.1109\/WACV51458.2022.00317"},{"key":"42_CR33","doi-asserted-by":"crossref","unstructured":"Wang, H., Li, Y., Wang, Y., Hu, H., Yang, M.H.: Collaborative distillation for ultra-resolution universal style transfer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1860\u20131869 (2020)","DOI":"10.1109\/CVPR42600.2020.00193"},{"key":"42_CR34","doi-asserted-by":"crossref","unstructured":"Wang, W., et al.: Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 568\u2013578 (2021)","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"42_CR35","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R., Gupta, A., He, K.: Non-local neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7794\u20137803 (2018)","DOI":"10.1109\/CVPR.2018.00813"},{"key":"42_CR36","unstructured":"Weissenborn, D., T\u00e4ckstr\u00f6m, O., Uszkoreit, J.: Scaling autoregressive video models. In: Proceedings of the International Conference on Learning Representations (ICLR) (2020)"},{"key":"42_CR37","unstructured":"Wu, C., et al.: Godiva: generating open-domain videos from natural descriptions. arXiv preprint arXiv:2104.14806 (2021)"},{"key":"42_CR38","unstructured":"Wu, C., et al.: N$$\\backslash $$\u201d UWA: visual synthesis pre-training for neural visual world creation. arXiv preprint arXiv:2111.12417 (2021)"},{"key":"42_CR39","unstructured":"Xu, K., et al.: Show, attend and tell: Neural image caption generation with visual attention. In: Proceedings of the International Conference on Machine Learning (ICML), pp. 2048\u20132057 (2015)"},{"key":"42_CR40","doi-asserted-by":"crossref","unstructured":"Xu, L., Du, Y., Zhang, Y.: An automatic framework for example-based virtual makeup. In: Proceedings of the IEEE International Conference on Image Processing (ICIP), pp. 3206\u20133210 (2013)","DOI":"10.1109\/ICIP.2013.6738660"},{"key":"42_CR41","doi-asserted-by":"crossref","unstructured":"Yuan, L., et al.: Tokens-to-token VIT: Training vision transformers from scratch on imageNet. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 558\u2013567 (2021)","DOI":"10.1109\/ICCV48922.2021.00060"},{"issue":"10","key":"42_CR42","doi-asserted-by":"publisher","first-page":"1499","DOI":"10.1109\/LSP.2016.2603342","volume":"23","author":"K Zhang","year":"2016","unstructured":"Zhang, K., Zhang, Z., Li, Z., Qiao, Y.: Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 23(10), 1499\u20131503 (2016)","journal-title":"IEEE Signal Process. Lett."},{"key":"42_CR43","doi-asserted-by":"crossref","unstructured":"Zhang, P., et al.: Multi-scale vision longformer: a new vision transformer for high-resolution image encoding. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 2998\u20133008 (2021)","DOI":"10.1109\/ICCV48922.2021.00299"},{"key":"42_CR44","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19787-1_42","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T07:59:08Z","timestamp":1728201548000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19787-1_42"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197864","9783031197871"],"references-count":44,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19787-1_42","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"21 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}