{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,4]],"date-time":"2026-02-04T17:12:15Z","timestamp":1770225135748,"version":"3.49.0"},"publisher-location":"Cham","reference-count":31,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030377304","type":"print"},{"value":"9783030377311","type":"electronic"}],"license":[{"start":{"date-parts":[[2019,12,24]],"date-time":"2019-12-24T00:00:00Z","timestamp":1577145600000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-37731-1_64","type":"book-chapter","created":{"date-parts":[[2019,12,27]],"date-time":"2019-12-27T06:02:51Z","timestamp":1577426571000},"page":"790-801","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Joint Sketch-Attribute Learning for Fine-Grained Face Synthesis"],"prefix":"10.1007","author":[{"given":"Binxin","family":"Yang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0478-7018","authenticated-orcid":false,"given":"Xuejin","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Richang","family":"Hong","sequence":"additional","affiliation":[]},{"given":"Zihan","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Yuhang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zheng-Jun","family":"Zha","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,12,24]]},"reference":[{"issue":"5","key":"64_CR1","first-page":"124:1","volume":"28","author":"T Chen","year":"2009","unstructured":"Chen, T., Cheng, M., Tan, P., Shamir, A., Hu, S.: Sketch2Photo: Internet image montage. ACM Trans. Graph. 28(5), 124:1\u2013124:10 (2009)","journal-title":"ACM Trans. Graph."},{"key":"64_CR2","doi-asserted-by":"crossref","unstructured":"Chen, W., Hays, J.: SketchyGAN: towards diverse and realistic sketch to image synthesis. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 9416\u20139425 (2018)","DOI":"10.1109\/CVPR.2018.00981"},{"key":"64_CR3","doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M., Kim, M., Ha, J., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: IEEE Computer Vision and Pattern Recognition, pp. 8789\u20138797 (2018)","DOI":"10.1109\/CVPR.2018.00916"},{"key":"64_CR4","unstructured":"Denton, E.L., Chintala, S., Szlam, A., Fergus, R.: Deep generative image models using a Laplacian pyramid of adversarial networks. In: Advances in Neural Information Processing Systems, pp. 1486\u20131494 (2015)"},{"key":"64_CR5","doi-asserted-by":"crossref","unstructured":"Dong, X., Yan, Y., Ouyang, W., Yang, Y.: Style aggregated network for facial landmark detection. In: IEEE Computer Vision and Pattern Recognition, pp. 379\u2013388 (2018)","DOI":"10.1109\/CVPR.2018.00047"},{"issue":"6","key":"64_CR6","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1109\/MCG.2011.67","volume":"31","author":"M Eitz","year":"2011","unstructured":"Eitz, M., Richter, R., Hildebrand, K., Boubekeur, T., Alexa, M.: Photosketcher: Interactive sketch-based image synthesis. IEEE Comput. Graph. Appl. 31(6), 56\u201366 (2011)","journal-title":"IEEE Comput. Graph. Appl."},{"key":"64_CR7","unstructured":"Goodfellow, I.J., et al.: Generative adversarial nets. In: Advances in Annual Conference on Neural Information Processing Systems 2014, pp. 2672\u20132680 (2014)"},{"issue":"11","key":"64_CR8","doi-asserted-by":"publisher","first-page":"5464","DOI":"10.1109\/TIP.2019.2916751","volume":"28","author":"Z He","year":"2019","unstructured":"He, Z., Zuo, W., Kan, M., Shan, S., Chen, X.: AttGAN: facial attribute editing by only changing what you want. IEEE Trans. Image Process. 28(11), 5464\u20135478 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"64_CR9","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 5967\u20135976 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"64_CR10","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of GANs for improved quality, stability, and variation. In: International Conference on Learning Representations (2018)"},{"key":"64_CR11","unstructured":"Kim, T., Cha, M., Kim, H., Lee, J.K., Kim, J.: Learning to discover cross-domain relations with generative adversarial networks. In: Proceedings of the 34th International Conference on Machine Learning, vol. 70, pp. 1857\u20131865 (2017)"},{"key":"64_CR12","doi-asserted-by":"crossref","unstructured":"Ledig, C., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 105\u2013114 (2017)","DOI":"10.1109\/CVPR.2017.19"},{"key":"64_CR13","doi-asserted-by":"crossref","unstructured":"Lee, D., Kim, J., Moon, W.J., Ye, J.C.: CollaGAN: collaborative GAN for missing image data imputation. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 2487\u20132496 (2019)","DOI":"10.1109\/CVPR.2019.00259"},{"key":"64_CR14","unstructured":"Liu, M.Y., Breuel, T., Kautz, J.: Unsupervised image-to-image translation networks. In: Advances in Neural Information Processing Systems, pp. 700\u2013708 (2017)"},{"key":"64_CR15","doi-asserted-by":"crossref","unstructured":"Liu, S., et al.: Face aging with contextual generative adversarial nets. In: ACM International Conference on Multimedia, pp. 82\u201390 (2018)","DOI":"10.1145\/3123266.3123431"},{"key":"64_CR16","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: IEEE International Conference on Computer Vision, pp. 3730\u20133738 (2015)","DOI":"10.1109\/ICCV.2015.425"},{"key":"64_CR17","doi-asserted-by":"crossref","unstructured":"Lu, Y., Tai, Y., Tang, C.: Attribute-guided face generation using conditional CycleGAN. In: ECCV, pp. 293\u2013308 (2018)","DOI":"10.1007\/978-3-030-01258-8_18"},{"key":"64_CR18","first-page":"3","volume-title":"MultiMedia Modeling","author":"Minho Park","year":"2018","unstructured":"Park, M., Kim, H.G., Ro, Y.M.: Photo-realistic facial emotion synthesis using multi-level critic networks with multi-level generative model. In: MultiMedia Modeling, pp. 3\u201315 (2019)"},{"key":"64_CR19","doi-asserted-by":"crossref","unstructured":"Pumarola, A., Agudo, A., Mart\u00ednez, A.M., Sanfeliu, A., Moreno-Noguer, F.: GANimation: anatomically-aware facial animation from a single image. In: ECCV, pp. 835\u2013851 (2018)","DOI":"10.1007\/978-3-030-01249-6_50"},{"key":"64_CR20","doi-asserted-by":"crossref","unstructured":"Wang, T., Liu, M., Zhu, J., Tao, A., Kautz, J., Catanzaro, B.: High-resolution image synthesis and semantic manipulation with conditional GANs. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 8798\u20138807 (2018)","DOI":"10.1109\/CVPR.2018.00917"},{"key":"64_CR21","doi-asserted-by":"crossref","unstructured":"Wang, X., Li, W., Mu, G., Huang, D., Wang, Y.: Facial expression synthesis by u-net conditional generative adversarial networks. In: ACM International Conference on Multimedia Retrieval, pp. 283\u2013290 (2018)","DOI":"10.1145\/3206025.3206068"},{"key":"64_CR22","doi-asserted-by":"crossref","unstructured":"Xiao, T., Hong, J., Ma, J.: ELEGANT: exchanging latent encodings with GAN for transferring multiple face attributes. In: ECCV, pp. 172\u2013187 (2018)","DOI":"10.1007\/978-3-030-01249-6_11"},{"key":"64_CR23","doi-asserted-by":"crossref","unstructured":"Xu, T., et al.: AttnGAN: fine-grained text to image generation with attentional generative adversarial networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1316\u20131324 (2018)","DOI":"10.1109\/CVPR.2018.00143"},{"key":"64_CR24","doi-asserted-by":"crossref","unstructured":"Yi, Z., Zhang, H., Tan, P., Gong, M.: DualGAN: unsupervised dual learning for image-to-image translation. In: IEEE International Conference on Computer Vision, pp. 2868\u20132876 (2017)","DOI":"10.1109\/ICCV.2017.310"},{"key":"64_CR25","doi-asserted-by":"crossref","unstructured":"Zhang, G., Kan, M., Shan, S., Chen, X.: Generative adversarial network with spatial attention for face attribute editing. In: ECCV, pp. 422\u2013437 (2018)","DOI":"10.1007\/978-3-030-01231-1_26"},{"key":"64_CR26","doi-asserted-by":"crossref","unstructured":"Zhang, H., Xu, T., Li, H.: StackGAN: text to photo-realistic image synthesis with stacked generative adversarial networks. In: IEEE International Conference on Computer Vision, pp. 5908\u20135916 (2017)","DOI":"10.1109\/ICCV.2017.629"},{"issue":"8","key":"64_CR27","doi-asserted-by":"publisher","first-page":"1947","DOI":"10.1109\/TPAMI.2018.2856256","volume":"41","author":"H Zhang","year":"2019","unstructured":"Zhang, H., et al.: StackGAN++: realistic image synthesis with stacked generative adversarial networks. IEEE Trans. Pattern Anal. Mach. Intell. 41(8), 1947\u20131962 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"64_CR28","doi-asserted-by":"crossref","unstructured":"Zhang, R., et al.: Style separation and synthesis via generative adversarial networks. In: ACM International Conference on Multimedia, pp. 183\u2013191 (2018)","DOI":"10.1145\/3240508.3240524"},{"key":"64_CR29","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Xie, Y., Yang, L.: Photographic text-to-image synthesis with a hierarchically-nested adversarial network. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 6199\u20136208 (2018)","DOI":"10.1109\/CVPR.2018.00649"},{"key":"64_CR30","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Deng, B., Huang, J., Lu, H., Hua, X.S.: Stylized adversarial autoencoder for image generation. In: ACM International Conference on Multimedia, pp. 244\u2013251 (2017)","DOI":"10.1145\/3123266.3123450"},{"key":"64_CR31","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: IEEE International Conference on Computer Vision, pp. 2242\u20132251 (2017)","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-37731-1_64","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,2,14]],"date-time":"2021-02-14T06:46:59Z","timestamp":1613285219000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-37731-1_64"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,12,24]]},"ISBN":["9783030377304","9783030377311"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-37731-1_64","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019,12,24]]},"assertion":[{"value":"24 December 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Daejeon","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Korea (Republic of)","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 January 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 January 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.mmm2020.kr\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"171","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"40","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"23% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Of the 171 submissions, 46 were accepted as poster papers; of the 49 special session paper submissions, 28 were accepted for oral presentation and 8 for poster presentation; 9 demo papers and 10 VBS papers were also accepted.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}