{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T16:46:20Z","timestamp":1774629980865,"version":"3.50.1"},"reference-count":48,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,7]]},"DOI":"10.1109\/ijcnn.2019.8851881","type":"proceedings-article","created":{"date-parts":[[2019,10,1]],"date-time":"2019-10-01T03:44:32Z","timestamp":1569901472000},"page":"1-8","source":"Crossref","is-referenced-by-count":119,"title":["Attention-Guided Generative Adversarial Networks for Unsupervised Image-to-Image Translation"],"prefix":"10.1109","author":[{"given":"Hao","family":"Tang","sequence":"first","affiliation":[]},{"given":"Dan","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Nicu","family":"Sebe","sequence":"additional","affiliation":[]},{"given":"Yan","family":"Yan","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1145\/3240508.3240704"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2019.8756586"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.723"},{"key":"ref32","article-title":"Learning what and where to draw","author":"reed","year":"2016","journal-title":"NIPS"},{"key":"ref31","article-title":"Generative adversarial text-to-image synthesis","author":"reed","year":"2016","journal-title":"ICML"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00263"},{"key":"ref37","article-title":"Unsupervised cross-domain image generation","author":"taigman","year":"2017","journal-title":"ICLRE"},{"key":"ref36","article-title":"Mask-aware photorealistic face attribute manipulation","author":"sun","year":"2018"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.578"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.135"},{"key":"ref10","doi-asserted-by":"crossref","DOI":"10.1145\/3072959.3073659","article-title":"Globally and locally consistent image completion","author":"iizuka","year":"2017","journal-title":"ACM TOG"},{"key":"ref40","article-title":"Multi-channel attention selection gan with cascaded semantic guidance for cross-view image translation","author":"tang","year":"2019","journal-title":"CVPR"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/IVMSPW.2018.8448850"},{"key":"ref13","article-title":"Learning to discover cross-domain relations with generative adversarial networks","author":"kim","year":"2017","journal-title":"ICML"},{"key":"ref14","article-title":"A method for stochastic optimization","author":"kingma","year":"2014","journal-title":"ICLRE"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1080\/02699930903485076"},{"key":"ref16","article-title":"Deep identity-aware transfer of facial attributes","author":"li","year":"2016"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.624"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01261-8_34"},{"key":"ref19","article-title":"Coupled generative adversarial networks","author":"liu","year":"2016","journal-title":"NIPS"},{"key":"ref28","article-title":"Invertible conditional gans for image editing","author":"perarnau","year":"2016","journal-title":"NIPS Workshop"},{"key":"ref4","article-title":"One-sided unsupervised domain mapping","author":"benaim","year":"2017","journal-title":"NIPS"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.82"},{"key":"ref3","article-title":"Openface: an open source facial behavior analysis toolkit","author":"baltru\u0161aitis","year":"2016","journal-title":"WACV"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01216-8_11"},{"key":"ref29","article-title":"Ganimation: Anatomically-aware facial animation from a single image","author":"pumarola","year":"2018","journal-title":"ECCV"},{"key":"ref5","article-title":"Large scale gan training for high fidelity natural image synthesis","author":"brock","year":"2019","journal-title":"ICLRE"},{"key":"ref8","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"NIPS"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00916"},{"key":"ref2","article-title":"Wasserstein gan","author":"arjovsky","year":"2017","journal-title":"ICML"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.267"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00122"},{"key":"ref46","article-title":"Genegan: Learning object transfiguration and attribute subspace from unpaired data","author":"zhou","year":"2017","journal-title":"BMVC"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.425"},{"key":"ref45","article-title":"Self-attention generative adversarial networks","author":"zhang","year":"2018"},{"key":"ref48","article-title":"Toward multimodal image-to-image translation","author":"zhu","year":"2017","journal-title":"NIPS"},{"key":"ref22","article-title":"Generating images from captions with attention","author":"mansimov","year":"2015","journal-title":"ICLRE"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref42","article-title":"Video-to-video synthesis","author":"wang","year":"2018","journal-title":"NeurIPS"},{"key":"ref24","article-title":"The ar face database","author":"martinez","year":"1998","journal-title":"CVC TR"},{"key":"ref41","article-title":"Dual generator generative adversarial networks for multi-domain image-to-image translation","author":"tang","year":"2018","journal-title":"ACCV"},{"key":"ref23","article-title":"Multi-class generative adversarial networks with the l2 loss function","author":"mao","year":"2016","journal-title":"CoRR"},{"key":"ref44","article-title":"A 3d facial expression database for facial behavior research","author":"yin","year":"2006","journal-title":"FG"},{"key":"ref26","article-title":"Conditional generative adversarial nets","author":"mirza","year":"2014"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.310"},{"key":"ref25","article-title":"Unsupervised attention-guided image to image translation","author":"mejjati","year":"2018","journal-title":"NeurIPS"}],"event":{"name":"2019 International Joint Conference on Neural Networks (IJCNN)","location":"Budapest, Hungary","start":{"date-parts":[[2019,7,14]]},"end":{"date-parts":[[2019,7,19]]}},"container-title":["2019 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8840768\/8851681\/08851881.pdf?arnumber=8851881","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,17]],"date-time":"2022-07-17T21:51:27Z","timestamp":1658094687000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8851881\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,7]]},"references-count":48,"URL":"https:\/\/doi.org\/10.1109\/ijcnn.2019.8851881","relation":{},"subject":[],"published":{"date-parts":[[2019,7]]}}}