{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T11:51:14Z","timestamp":1774957874361,"version":"3.50.1"},"reference-count":53,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61773302"],"award-info":[{"award-number":["61773302"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61906142"],"award-info":[{"award-number":["61906142"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005320","name":"Xidian University","doi-asserted-by":"publisher","award":["10221150004"],"award-info":[{"award-number":["10221150004"]}],"id":[{"id":"10.13039\/501100005320","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2019]]},"DOI":"10.1109\/access.2019.2939654","type":"journal-article","created":{"date-parts":[[2019,9,5]],"date-time":"2019-09-05T19:57:47Z","timestamp":1567713467000},"page":"126651-126661","source":"Crossref","is-referenced-by-count":27,"title":["Consistent Embedded GAN for Image-to-Image Translation"],"prefix":"10.1109","volume":"7","author":[{"given":"Feng","family":"Xiong","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2998-6582","authenticated-orcid":false,"given":"Qianqian","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Quanxue","family":"Gao","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.168"},{"key":"ref38","first-page":"700","article-title":"Unsupervised image-to-image translation networks","author":"liu","year":"2017","journal-title":"Proc NIPS"},{"key":"ref33","first-page":"3483","article-title":"Learning structured output representation using deep conditional generative models","author":"sohn","year":"2015","journal-title":"Proc NIPS"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00917"},{"key":"ref31","article-title":"Conditional generative adversarial nets","author":"mirza","year":"2014","journal-title":"arXiv 1411 1784"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.202"},{"key":"ref37","first-page":"752","article-title":"One-sided unsupervised domain mapping","author":"benaim","year":"2017","journal-title":"Proc NIPS"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.18"},{"key":"ref35","first-page":"1","article-title":"Unsupervised cross-domain image generation","author":"taigman","year":"2017","journal-title":"Proc ICLR"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.241"},{"key":"ref28","first-page":"1","article-title":"LR-GAN: Layered recursive generative adversarial networks for image generation","author":"yang","year":"2017","journal-title":"Proc ICLR"},{"key":"ref27","first-page":"658","article-title":"Generating images with perceptual similarity metrics based on deep networks","author":"dosovitskiy","year":"2016","journal-title":"Proc NIPS"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00916"},{"key":"ref2","first-page":"465","article-title":"Toward multimodal image-to-image translation","author":"zhu","year":"2017","journal-title":"Proc NIPS"},{"key":"ref1","first-page":"649","article-title":"Colorful image colorization","author":"zhang","year":"2016","journal-title":"Proc ECCV"},{"key":"ref20","first-page":"1","article-title":"Deep multi-scale video prediction beyond mean square error","author":"mathieu","year":"2016","journal-title":"Proc ICIP"},{"key":"ref22","first-page":"179","article-title":"Multimodal unsupervised image-to-image translation","author":"huang","year":"2018","journal-title":"Proc ECCV"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00579"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.304"},{"key":"ref23","first-page":"2234","article-title":"Improved techniques for training GANs","author":"salimans","year":"2016","journal-title":"Proc NIPS"},{"key":"ref26","first-page":"3308","article-title":"VEEGAN: Reducing mode collapse in gans using implicit variational learning","author":"srivastava","year":"2017","journal-title":"Proc NIPS"},{"key":"ref25","first-page":"1","article-title":"Wasserstein auto-encoders","author":"tolstikhin","year":"2018","journal-title":"Proc ICLR"},{"key":"ref50","first-page":"364","article-title":"Spatial pattern templates for recognition of objects with regular structure","author":"tyle?ek","year":"2013","journal-title":"Proc GCPR"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1016\/0047-259X(82)90077-X"},{"key":"ref52","first-page":"1097","article-title":"ImageNet classification with deep convolutional neural networks","author":"krizhevsky","year":"2012","journal-title":"Proc NIPS"},{"key":"ref10","first-page":"184","article-title":"Learning a deep convolutional network for image super-resolution","author":"dong","year":"2014","journal-title":"Proc ECCV"},{"key":"ref11","first-page":"5047","article-title":"Disentangling factors of variation in deep representation using adversarial training","author":"mathieu","year":"2016","journal-title":"Proc NIPS"},{"key":"ref40","first-page":"1","article-title":"PixeLNN: Example-based image synthesis","author":"bansal","year":"2018","journal-title":"Proc ICLR"},{"key":"ref12","first-page":"1486","article-title":"Deep generative image models using a laplacian pyramid of adversarial networks","author":"denton","year":"2015","journal-title":"Proc NIPS"},{"key":"ref13","first-page":"97","article-title":"Unsupervised representation learning with deep convolutional generative adversarial networks","author":"radford","year":"2016","journal-title":"Proc ICIP"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.310"},{"key":"ref15","first-page":"1857","article-title":"Learning to discover cross-domain relations with generative adversarial networks","author":"kim","year":"2017","journal-title":"Proc ICML"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.723"},{"key":"ref18","first-page":"469","article-title":"Coupled generative adversarial networks","author":"liu","year":"2016","journal-title":"Proc NIPS"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2017.2753232"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00986"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref6","first-page":"577","article-title":"Learning representations for automatic colorization","author":"larsson","year":"2016","journal-title":"Proc ECCV"},{"key":"ref5","doi-asserted-by":"crossref","first-page":"110","DOI":"10.1145\/2897824.2925974","article-title":"Let there be color!: Joint end-to-end learning of global and local image priors for automatic image colorization with simultaneous classification","volume":"35","author":"iizuka","year":"2016","journal-title":"ACM Trans Graph"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.434"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.278"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.350"},{"key":"ref9","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc NIPS"},{"key":"ref46","first-page":"597","article-title":"Generative visual manipulation on the natural image manifold","author":"zhu","year":"2016","journal-title":"Proc ECCV"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.32"},{"key":"ref48","doi-asserted-by":"crossref","first-page":"149","DOI":"10.1145\/2601097.2601101","article-title":"Transient attributes for high-level understanding and editing of outdoor scenes","volume":"33","author":"laffont","year":"2014","journal-title":"ACM Trans Graph"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-017-1004-z"},{"key":"ref42","first-page":"1","article-title":"Automatic differentiation in pytorch","author":"paszke","year":"2017","journal-title":"Proc ICML"},{"key":"ref41","first-page":"82","article-title":"Learning a probabilistic latent space of object shapes via 3D generative-adversarial modeling","author":"wu","year":"2016","journal-title":"Proc NIPS"},{"key":"ref44","first-page":"1","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"2015","journal-title":"Proc ICIP"},{"key":"ref43","first-page":"234","article-title":"U-net: Convolutional networks for biomedical image segmentation","author":"ronneberger","year":"2015","journal-title":"Proc MICCAI"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/8600701\/08825805.pdf?arnumber=8825805","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,27]],"date-time":"2022-01-27T09:40:18Z","timestamp":1643276418000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8825805\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"references-count":53,"URL":"https:\/\/doi.org\/10.1109\/access.2019.2939654","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019]]}}}