{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,20]],"date-time":"2025-10-20T10:28:09Z","timestamp":1760956089537,"version":"build-2065373602"},"reference-count":47,"publisher":"MDPI AG","issue":"5","license":[{"start":{"date-parts":[[2021,5,16]],"date-time":"2021-05-16T00:00:00Z","timestamp":1621123200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Entropy"],"abstract":"<jats:p>In the context of social media, large amounts of headshot photos are taken everyday. Unfortunately, in addition to laborious editing and modification, creating a visually compelling photographic masterpiece for sharing requires advanced professional skills, which are difficult for ordinary Internet users. Though there are many algorithms automatically and globally transferring the style from one image to another, they fail to respect the semantics of the scene and are unable to allow users to merely transfer the attributes of one or two face organs in the foreground region leaving the background region unchanged. To overcome this problem, we developed a novel framework for semantically meaningful local face attribute transfer, which can flexibly transfer the local attribute of a face organ from the reference image to a semantically equivalent organ in the input image, while preserving the background. Our method involves warping the reference photo to match the shape, pose, location, and expression of the input image. The fusion of the warped reference image and input image is then taken as the initialized image for a neural style transfer algorithm. Our method achieves better performance in terms of inception score (3.81) and Fr\u00e9chet inception distance (80.31), which is about 10% higher than those of competitors, indicating that our framework is capable of producing high-quality and photorealistic attribute transfer results. Both theoretical findings and experimental results are provided to demonstrate the efficacy of the proposed framework, reveal its superiority over other state-of-the-art alternatives.<\/jats:p>","DOI":"10.3390\/e23050615","type":"journal-article","created":{"date-parts":[[2021,5,16]],"date-time":"2021-05-16T23:17:16Z","timestamp":1621207036000},"page":"615","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["A Foreground-Aware Framework for Local Face Attribute Transfer"],"prefix":"10.3390","volume":"23","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2011-5337","authenticated-orcid":false,"given":"Yuanbin","family":"Fu","sequence":"first","affiliation":[{"name":"College of Intelligence and Computing, Tianjin University, Tianjin 300350, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3264-3265","authenticated-orcid":false,"given":"Jiayi","family":"Ma","sequence":"additional","affiliation":[{"name":"Electronic Information School, Wuhan University, Wuhan 430072, China"}]},{"given":"Xiaojie","family":"Guo","sequence":"additional","affiliation":[{"name":"College of Intelligence and Computing, Tianjin University, Tianjin 300350, China"}]}],"member":"1968","published-online":{"date-parts":[[2021,5,16]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"3365","DOI":"10.1109\/TVCG.2019.2921336","article-title":"Neural style transfer: A review","volume":"26","author":"Jing","year":"2019","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"148:1","DOI":"10.1145\/2601097.2601137","article-title":"Style transfer for headshot portraits","volume":"33","author":"Shih","year":"2014","journal-title":"TOG"},{"doi-asserted-by":"crossref","unstructured":"Gatys, L.A., Ecker, A.S., and Bethge, M. (2016, January 27\u201330). Image style transfer using convolutional neural networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","key":"ref_3","DOI":"10.1109\/CVPR.2016.265"},{"unstructured":"Leibe, B., Matas, J., Sebe, N., and Welling, M. (2016). Perceptual losses for real-time style transfer and super-resolution. European Conference on Computer Vision, Springer.","key":"ref_4"},{"doi-asserted-by":"crossref","unstructured":"Chen, D., Yuan, L., Liao, J., Yu, N., and Hua, G. (2017, January 21\u201326). Stylebank: An explicit representation for neural image style transfer. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","key":"ref_5","DOI":"10.1109\/CVPR.2017.296"},{"doi-asserted-by":"crossref","unstructured":"Li, Y., Wang, N., Liu, J., and Hou, X. (2017). Demystifying neural style transfer. arXiv.","key":"ref_6","DOI":"10.24963\/ijcai.2017\/310"},{"unstructured":"Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., and Yang, M. (2017, January 4\u20139). Universal style transfer via feature transforms. Proceedings of the NIPS, Long Beach, CA, USA.","key":"ref_7"},{"doi-asserted-by":"crossref","unstructured":"Sheng, L., Lin, Z., Shao, J., and Wang, X. (2018, January 18\u201323). Avatar-net: Multi-scale zero-shot style transfer by feature decoration. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","key":"ref_8","DOI":"10.1109\/CVPR.2018.00860"},{"doi-asserted-by":"crossref","unstructured":"Shen, F., Yan, S., and Zeng, G. (2018, January 18\u201323). Neural style transfer via meta networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","key":"ref_9","DOI":"10.1109\/CVPR.2018.00841"},{"doi-asserted-by":"crossref","unstructured":"Huang, H., Wang, H., Luo, W., Ma, L., Jiang, W., Zhu, X., Li, Z., and Liu, W. (2017, January 21\u201326). Real-time neural style transfer for videos. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","key":"ref_10","DOI":"10.1109\/CVPR.2017.745"},{"doi-asserted-by":"crossref","unstructured":"Luan, F., Paris, S., Shechtman, E., and Bala, K. (2017, January 21\u201326). Deep photo style transfer. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","key":"ref_11","DOI":"10.1109\/CVPR.2017.740"},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"679","DOI":"10.1145\/1015706.1015779","article-title":"Non-photorealistic camera: Depth edge detection and stylized rendering using multi-flash imaging","volume":"23","author":"Raskar","year":"2004","journal-title":"TOG"},{"doi-asserted-by":"crossref","unstructured":"Liu, X., Cheng, M., Lai, Y., and Rosin, P.L. (2017, January 29\u201330). Depth-aware neural style transfer. Proceedings of the NPAR, Los Angeles, CA, USA.","key":"ref_13","DOI":"10.1145\/3092919.3092924"},{"doi-asserted-by":"crossref","unstructured":"Hertzmann, A., Jacobs, C.E., Oliver, N., Curless, B., and Salesin, D. (2001, January 12\u201317). Image analogies. Proceedings of the SIGGRAPH, Los Angeles, CA, USA.","key":"ref_14","DOI":"10.1145\/383259.383295"},{"doi-asserted-by":"crossref","unstructured":"Men, Y., Lian, Z., Tang, Y., and Xiao, J. (2018, January 18\u201323). A common framework for interactive texture transfer. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","key":"ref_15","DOI":"10.1109\/CVPR.2018.00665"},{"unstructured":"Simonyan, K., and Zisserman, A. (2015, January 7\u20139). Very deep convolutional networks for large-scale image recognition. Proceedings of the ICLR, San Diego, CA, USA.","key":"ref_16"},{"doi-asserted-by":"crossref","unstructured":"Gu, S., Chen, C., Liao, J., and Yuan, L. (2018, January 18\u201323). Arbitrary style transfer with deep feature reshuffle. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","key":"ref_17","DOI":"10.1109\/CVPR.2018.00858"},{"doi-asserted-by":"crossref","unstructured":"Chang, H., Lu, J., Yu, F., and Finkelstein, A. (2018, January 18\u201323). Pairedcyclegan: Asymmetric style transfer for applying and removing makeup. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","key":"ref_18","DOI":"10.1109\/CVPR.2018.00012"},{"doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M., Kim, M., Ha, J., Kim, S., and Choo, J. (2018, January 18\u201323). Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","key":"ref_19","DOI":"10.1109\/CVPR.2018.00916"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"244:1","DOI":"10.1145\/3272127.3275046","article-title":"CariGANs: Unpaired photo-to-caricature translation","volume":"37","author":"Cao","year":"2018","journal-title":"TOG"},{"doi-asserted-by":"crossref","unstructured":"Kemelmacher-Shlizerman, I., Suwajanakorn, S., and Seitz, S.M. (2014, January 23\u201328). Illumination-aware age progression. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA.","key":"ref_21","DOI":"10.1109\/CVPR.2014.426"},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"641","DOI":"10.1111\/1467-8659.t01-1-00712","article-title":"Reanimating faces in images and video","volume":"22","author":"Blanz","year":"2003","journal-title":"Comput. Graph. Forum"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"1968","DOI":"10.1109\/TPAMI.2008.244","article-title":"Face relighting from a single image under arbitrary unknown lighting conditions","volume":"31","author":"Wang","year":"2009","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"60","DOI":"10.1145\/2010324.1964955","article-title":"Expression flow for 3D-aware face component transfer","volume":"30","author":"Yang","year":"2011","journal-title":"TOG"},{"doi-asserted-by":"crossref","unstructured":"Korshunova, I., Shi, W., Dambre, J., and Theis, L. (2017, January 21\u201326). Fast face-swap using convolutional neural networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","key":"ref_25","DOI":"10.1109\/ICCV.2017.397"},{"doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J., Zhou, T., and Efros, A.A. (2017, January 21\u201326). Image-to-image translation with conditional adversarial networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","key":"ref_26","DOI":"10.1109\/CVPR.2017.632"},{"doi-asserted-by":"crossref","unstructured":"Zhu, J., Park, T., Isola, P., and Efros, A.A. (2017, January 22\u201329). Unpaired image-to-image translation using cycle-consistent adversarial networks. Proceedings of the ICCV, Venice, Italy.","key":"ref_27","DOI":"10.1109\/ICCV.2017.244"},{"doi-asserted-by":"crossref","unstructured":"Yi, Z., Zhang, H.R., Tan, P., and Gong, M. (2017, January 22\u201329). Dualgan: Unsupervised dual learning for image-to-image translation. Proceedings of the ICCV, Venice, Italy.","key":"ref_28","DOI":"10.1109\/ICCV.2017.310"},{"unstructured":"Kim, T., Cha, M., Kim, H., Lee, J.K., and Kim, J. (2017, January 6\u201311). Learning to discover cross-domain relations with generative adversarial networks. Proceedings of the ICML, Sydney, NSW, Australia.","key":"ref_29"},{"doi-asserted-by":"crossref","unstructured":"Choi, Y., Uh, Y., Yoo, J., and Ha, J. (2020, January 13\u201319). Stargan v2: Diverse image synthesis for multiple domains. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA.","key":"ref_30","DOI":"10.1109\/CVPR42600.2020.00821"},{"doi-asserted-by":"crossref","unstructured":"Hui, L., Li, X., Chen, J., He, H., and Yang, J. (2018, January 20\u201324). Unsupervised multi-domain image translation with domain-specific encoders\/decoders. Proceedings of the ICPR, Beijing, China.","key":"ref_31","DOI":"10.1109\/ICPR.2018.8545169"},{"unstructured":"Liu, M., Huang, X., Mallya, A., Karras, T., Aila, T., Lehtinen, J., and Kautz, J. (November, January 27). Few-shot unsupervised image-to-image translation. Proceedings of the ICCV, Seoul, Korea.","key":"ref_32"},{"doi-asserted-by":"crossref","unstructured":"Chen, Y., Lin, H., Shu, M., Li, R., Tao, X., Shen, X., Ye, Y., and Jia, J. (2018, January 18\u201323). Facelet-bank for fast portrait manipulation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","key":"ref_33","DOI":"10.1109\/CVPR.2018.00373"},{"doi-asserted-by":"crossref","unstructured":"Chen, Y., Shen, X., Lin, Z., Lu, X., Pao, I., and Jia, J. (2019, January 16\u201320). Semantic component decomposition for face attribute manipulation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","key":"ref_34","DOI":"10.1109\/CVPR.2019.01009"},{"unstructured":"Sabour, S., Frosst, N., and Hinton, G.E. (2017, January 4\u20139). Dynamic routing between capsules. Proceedings of the NIPS, Long Beach, CA, USA.","key":"ref_35"},{"doi-asserted-by":"crossref","unstructured":"Breitenstein, M.D., K\u00fcttel, D., Weise, T., Gool, L.V., and Pfister, H. (2008, January 24\u201326). Real-time face pose estimation from single range images. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Anchorage, AK, USA.","key":"ref_36","DOI":"10.1109\/CVPR.2008.4587807"},{"unstructured":"Zhu, X., and Ramanan, D. (2012, January 16\u201321). Face detection, pose estimation, and landmark localization in the wild. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA.","key":"ref_37"},{"doi-asserted-by":"crossref","unstructured":"Bulat, A., and Tzimiropoulos, G. (2017, January 21\u201326). Binarized convolutional landmark localizers for human pose estimation and face alignment with limited resources. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","key":"ref_38","DOI":"10.1109\/ICCV.2017.400"},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"121549","DOI":"10.1109\/ACCESS.2020.3006958","article-title":"Facial micro-expression recognition using two-dimensional landmark feature maps","volume":"8","author":"Choi","year":"2020","journal-title":"IEEE Access"},{"doi-asserted-by":"crossref","unstructured":"Kim, J.H., Poulose, A., and Han, D.S. (2021). The extensive usage of the facial image threshing machine for facial emotion recognition performance. Sensors, 21.","key":"ref_40","DOI":"10.3390\/s21062026"},{"doi-asserted-by":"crossref","unstructured":"Kumar, A., Marks, T.K., Mou, W., Wang, Y., Jones, M., Cherian, A., Koike-Akino, T., Liu, X., and Feng, C. (2020, January 13\u201319). LUVLi face alignment: Estimating landmarks\u2019 location, uncertainty, and visibility likelihood. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA.","key":"ref_41","DOI":"10.1109\/CVPR42600.2020.00826"},{"doi-asserted-by":"crossref","unstructured":"Kazemi, V., and Sullivan, J. (2014, January 23\u201328). One millisecond face alignment with an ensemble of regression trees. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA.","key":"ref_42","DOI":"10.1109\/CVPR.2014.241"},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"567","DOI":"10.1109\/34.24792","article-title":"Principal warps: Thin-plate splines and the decomposition of deformations","volume":"11","author":"Bookstein","year":"1989","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"unstructured":"Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X., and Chen, X. (2016, January 5\u201310). Improved techniques for training GANs. Proceedings of the NIPS, Barcelona, Spain.","key":"ref_44"},{"doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., and Rabinovich, A. (2015, January 7\u201312). Going deeper with convolutions. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA.","key":"ref_45","DOI":"10.1109\/CVPR.2015.7298594"},{"unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., and Hochreiter, S. (2017, January 4\u20139). GANs trained by a two time-scale update rule converge to a local nash equilibrium. Proceedings of the NIPS, Long Beach, CA, USA.","key":"ref_46"},{"doi-asserted-by":"crossref","unstructured":"Smith, B.M., Zhang, L., Brandt, J., Lin, Z., and Yang, J. (2013, January 23\u201328). Exemplar-based face parsing. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA.","key":"ref_47","DOI":"10.1109\/CVPR.2013.447"}],"container-title":["Entropy"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1099-4300\/23\/5\/615\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T06:02:10Z","timestamp":1760162530000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1099-4300\/23\/5\/615"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,5,16]]},"references-count":47,"journal-issue":{"issue":"5","published-online":{"date-parts":[[2021,5]]}},"alternative-id":["e23050615"],"URL":"https:\/\/doi.org\/10.3390\/e23050615","relation":{},"ISSN":["1099-4300"],"issn-type":[{"type":"electronic","value":"1099-4300"}],"subject":[],"published":{"date-parts":[[2021,5,16]]}}}