{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,6]],"date-time":"2026-04-06T07:22:31Z","timestamp":1775460151505,"version":"3.50.1"},"reference-count":258,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U1908209"],"award-info":[{"award-number":["U1908209"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61632001"],"award-info":[{"award-number":["61632001"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62021001"],"award-info":[{"award-number":["62021001"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"National Key Research and Development Program of China","award":["2018AAA0101400"],"award-info":[{"award-number":["2018AAA0101400"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Multimedia"],"published-print":{"date-parts":[[2022]]},"DOI":"10.1109\/tmm.2021.3109419","type":"journal-article","created":{"date-parts":[[2021,9,3]],"date-time":"2021-09-03T20:02:55Z","timestamp":1630699375000},"page":"3859-3881","source":"Crossref","is-referenced-by-count":291,"title":["Image-to-Image Translation: Methods and Applications"],"prefix":"10.1109","volume":"24","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2084-5177","authenticated-orcid":false,"given":"Yingxue","family":"Pang","sequence":"first","affiliation":[{"name":"Department of Electronic Engineer and Information Science, University of Science and Technology of China, Hefei, Anhui, China"}]},{"given":"Jianxin","family":"Lin","sequence":"additional","affiliation":[{"name":"School of Computer Science and Electronic Engineering, Hunan University, Changsha, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9095-0776","authenticated-orcid":false,"given":"Tao","family":"Qin","sequence":"additional","affiliation":[{"name":"Microsoft Research Asia, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8525-5066","authenticated-orcid":false,"given":"Zhibo","family":"Chen","sequence":"additional","affiliation":[{"name":"Department of Electronic Engineer and Information Science, University of Science and Technology of China, Hefei, Anhui, China"}]}],"member":"263","reference":[{"key":"ref170","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00259"},{"key":"ref172","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00490"},{"key":"ref171","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2019\/413"},{"key":"ref174","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2019\/345"},{"key":"ref173","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00028"},{"key":"ref176","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58529-7_24"},{"key":"ref175","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00602"},{"key":"ref178","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2019.2950198"},{"key":"ref177","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01249-6_50"},{"key":"ref168","first-page":"5914","article-title":"RELGAN: Multi-domain image-to-image translation via relative attributes","author":"wu","year":"0","journal-title":"Proc IEEE\/CVF Int Conf Comput Vis"},{"key":"ref169","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00379"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01342"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00811"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01102"},{"key":"ref32","first-page":"2590","article-title":"A unified feature disentangler for multi-domain image translation and manipulation","author":"liu","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref31","article-title":"DIDA: Disentangled synthesis for domain adaptation","author":"cao","year":"2018","journal-title":"CoRR"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00473"},{"key":"ref37","article-title":"U-GAT-IT: Unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation","author":"kim","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00986"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2019.04.032"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.3233\/FI-2019-1834"},{"key":"ref181","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-019-01284-z"},{"key":"ref180","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"ref185","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00451"},{"key":"ref184","article-title":"Attribute guided unpaired image-to-image translation with semi-supervised learning","author":"li","year":"2019"},{"key":"ref183","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58580-8_23"},{"key":"ref182","article-title":"GMM-unit: Unsupervised multi-domain and multi-modal image-to-image translation via attribute gaussian mixture modeling","author":"liu","year":"2020"},{"key":"ref189","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2019.2922854"},{"key":"ref188","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.32"},{"key":"ref187","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.167"},{"key":"ref186","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.425"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00113"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00584"},{"key":"ref179","first-page":"2994","article-title":"Multi-mapping image-to-image translation via learning disentanglement","author":"yu","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2019.2938347"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.467"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1145\/3450626.3459783"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01317"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1145\/3197517.3201365"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2017.32"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00938"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00824"},{"key":"ref50","first-page":"1747","article-title":"Pixel recurrent neural networks","author":"oord","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref51","first-page":"864","article-title":"Pixelsnail: An improved autoregressive generative model","author":"chen","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref154","article-title":"Nips 2016 tutorial: Generative adversarial networks","author":"goodfellow","year":"2017"},{"key":"ref153","doi-asserted-by":"publisher","DOI":"10.1145\/383259.383295"},{"key":"ref156","first-page":"1558","article-title":"Autoencoding beyond pixels using a learned similarity metric","author":"larsen","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref155","doi-asserted-by":"crossref","first-page":"504","DOI":"10.1126\/science.1127647","article-title":"Reducing the dimensionality of data with neural networks","volume":"313","author":"hinton","year":"2006","journal-title":"Science"},{"key":"ref150","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00049"},{"key":"ref152","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00131"},{"key":"ref151","first-page":"4077","article-title":"Prototypical networks for few-shot learning","author":"snell","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref146","first-page":"3581","article-title":"Semi-supervised learning with deep generative models","author":"kingma","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref147","first-page":"3546","article-title":"Semi-supervised learning with ladder networks","author":"rasmus","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref148","first-page":"5049","article-title":"Mixmatch: A holistic approach to semi-supervised learning","author":"berthelot","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref149","first-page":"2365","article-title":"MetaGAN: An adversarial approach to few-shot learning","author":"zhang","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref59","first-page":"131","article-title":"Approximate inference for deep latent gaussian mixtures","volume":"2","author":"nalisnick","year":"0","journal-title":"Proc NIPS Workshop Bayesian Deep Learn"},{"key":"ref58","first-page":"881","article-title":"Made: Masked autoencoder for distribution estimation","author":"germain","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref57","first-page":"29","article-title":"The neural autoregressive distribution estimator","author":"larochelle","year":"0","journal-title":"Proc 14th Int Conf Artif Intell Statist"},{"key":"ref56","first-page":"1278","article-title":"Stochastic backpropagation and approximate inference in deep generative models","author":"rezende","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref55","first-page":"1","article-title":"Auto-encoding variational bayes","volume":"1050","author":"kingma","year":"2014","journal-title":"Proc 2nd Int Conf Learn Representations ICLR 2014"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1995.7.5.889"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00326"},{"key":"ref52","article-title":"Nice: Non-linear independent components estimation","author":"dinh","year":"2014"},{"key":"ref40","article-title":"A tutorial on energy-based learning","volume":"1","author":"lecun","year":"2006","journal-title":"Predicting Structured Data"},{"key":"ref167","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2019.2916751"},{"key":"ref166","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00916"},{"key":"ref165","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01264-9_10"},{"key":"ref164","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2018.8545169"},{"key":"ref163","article-title":"Representation learning with contrastive predictive coding","author":"oord","year":"2018"},{"key":"ref162","first-page":"4414","article-title":"Unsupervised learning of disentangled representations from video","author":"denton","year":"0","journal-title":"Proc Adv Neural Inf Process Syst 30"},{"key":"ref161","first-page":"2649","article-title":"Disentangling by factorising","author":"kim","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref160","article-title":"Beta-VAE: Learning basic visual concepts with a constrained variational framework","author":"higgins","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00244"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00369"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00559"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00515"},{"key":"ref159","article-title":"Adversarially learned inference","volume":"1050","author":"dumoulin","year":"2017","journal-title":"Proc 5th Int Conf Learn Representations ICLR 2017"},{"key":"ref8","article-title":"MRI cross-modality neuroimage-to-neuroimage translation","author":"yang","year":"2018"},{"key":"ref49","first-page":"4790","article-title":"Conditional image generation with pixelcnn decoders","author":"oord","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00789"},{"key":"ref157","first-page":"2172","article-title":"InfoGAN: Interpretable representation learning by information maximizing generative adversarial nets","author":"chen","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref158","article-title":"Adversarial feature learning","author":"donahue","year":"2017","journal-title":"Proc 5th Int Conf Learn Representations ICLR 2017"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2019.01.115"},{"key":"ref46","first-page":"448","article-title":"Deep boltzmann machines","author":"salakhutdinov","year":"0","journal-title":"Proc Artif Intell Statist"},{"key":"ref45","first-page":"5929","article-title":"Deep generative models for distribution-preserving lossy compression","author":"tschannen","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1162\/neco.2006.18.7.1527"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1145\/1273496.1273596"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/ISACV.2018.8354080"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1080\/02564602.2014.987328"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.728"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01216-8_25"},{"key":"ref73","article-title":"From variational to deterministic autoencoders","author":"ghosh","year":"0","journal-title":"Int Conf Learn Representations"},{"key":"ref72","article-title":"Spectral normalization for generative adversarial networks","author":"miyato","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref71","article-title":"The relativistic discriminator: A key element missing from standard GAN","author":"jolicoeur-martineau","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref70","article-title":"BeGAN: Boundary equilibrium generative adversarial networks","author":"berthelot","year":"2017"},{"key":"ref76","article-title":"Exemplar guided unsupervised image-to-image translation with semantic consistency","author":"ma","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref77","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00820"},{"key":"ref74","first-page":"700","article-title":"Unsupervised image-to-image translation networks","author":"liu","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref75","first-page":"35","article-title":"Diverse image-to-image translation via disentangled representations","author":"lee","year":"0","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref78","first-page":"10 348","article-title":"Unsupervised image-to-image translation using domain-specific variational information bound","author":"kazemi","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref79","article-title":"ZSTGAN: An adversarial approach for unsupervised zero-shot image-to-image translation","author":"lin","year":"2019"},{"key":"ref60","first-page":"1530","article-title":"Variational inference with normalizing flows","author":"rezende","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref62","article-title":"Wasserstein auto-encoders","author":"tolstikhin","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref61","first-page":"1214","article-title":"Vae with a vampprior","author":"tomczak","year":"0","journal-title":"Proc Int Conf Artif Intell Statist"},{"key":"ref63","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref64","first-page":"214","article-title":"Wasserstein generative adversarial networks","author":"arjovsky","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.304"},{"key":"ref66","article-title":"Unsupervised representation learning with deep convolutional generative adversarial networks","author":"radford","year":"0","journal-title":"Proc 4th Int Conf Learn Representations"},{"key":"ref67","article-title":"Conditional generative adversarial nets","author":"mirza","year":"2014"},{"key":"ref68","first-page":"5767","article-title":"Improved training of wasserstein GANs","author":"gulrajani","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref69","article-title":"Energy-based generative adversarial networks","author":"zhao","year":"0","journal-title":"Proc 5th Int Conf Learn Representations"},{"key":"ref197","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2020.3015015"},{"key":"ref198","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00748"},{"key":"ref199","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00012"},{"key":"ref193","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00923"},{"key":"ref194","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00912"},{"key":"ref195","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2020.3001536"},{"key":"ref196","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00981"},{"key":"ref95","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00467"},{"key":"ref94","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.308"},{"key":"ref190","doi-asserted-by":"publisher","DOI":"10.1145\/3123266.3123277"},{"key":"ref93","article-title":"Demystifying MMD GANs","author":"bi?kowski","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref191","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00359"},{"key":"ref92","first-page":"6626","article-title":"GANs trained by a two time-scale update rule converge to a local nash equilibrium","author":"heusel","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref192","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00018"},{"key":"ref91","first-page":"694","article-title":"Perceptual losses for real-time style transfer and super-resolution","author":"johnson","year":"0","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref90","first-page":"172","article-title":"Multimodal unsupervised image-to-image translation","author":"huang","year":"0","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref98","first-page":"7176","article-title":"Reliable fidelity and diversity metrics for generative models","author":"naeem","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref99","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00917"},{"key":"ref96","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"ref97","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref82","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1989.1.4.541"},{"key":"ref81","doi-asserted-by":"publisher","DOI":"10.1109\/72.159058"},{"key":"ref84","first-page":"26","article-title":"Lecture 6.5-RmsProp: Divide the gradient by a running average of its recent magnitude","volume":"4","author":"tieleman","year":"0","journal-title":"COURSERA Neural Netw Mach Learn"},{"key":"ref83","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"0","journal-title":"Proc 3rd Int Conf Learn Representations"},{"key":"ref80","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58548-8_2"},{"key":"ref89","article-title":"Pose guided person image generation","volume":"30","author":"ma","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref85","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46487-9_40"},{"key":"ref86","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01246-5_47"},{"key":"ref87","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2003.819861"},{"key":"ref88","first-page":"2234","article-title":"Improved techniques for training gans","author":"salimans","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref200","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2020.2975961"},{"key":"ref101","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00252"},{"key":"ref100","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00911"},{"key":"ref209","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00173"},{"key":"ref203","first-page":"2611","article-title":"MangaGAN: Unpaired photo-to-manga translation based on the methodology of manga drawing","volume":"35","author":"su","year":"0","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"ref204","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i01.5405"},{"key":"ref201","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01228-1_8"},{"key":"ref202","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00552"},{"key":"ref207","article-title":"Powers of layers for image-to-image translation","author":"touvron","year":"2020"},{"key":"ref208","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2019.2920407"},{"key":"ref205","doi-asserted-by":"publisher","DOI":"10.1016\/j.compmedimag.2019.101684"},{"key":"ref206","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-33391-1_1"},{"key":"ref211","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2019.00253"},{"key":"ref210","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33019332"},{"key":"ref212","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00127"},{"key":"ref213","doi-asserted-by":"publisher","DOI":"10.1016\/j.ifacol.2019.12.287"},{"key":"ref214","doi-asserted-by":"publisher","DOI":"10.1109\/APSIPAASC47483.2019.9023296"},{"key":"ref215","doi-asserted-by":"publisher","DOI":"10.1007\/s12555-019-0689-x"},{"key":"ref216","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00854"},{"key":"ref217","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-11021-5_10"},{"key":"ref218","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00897"},{"key":"ref219","first-page":"353","article-title":"Unsupervised class-specific deblurring","author":"nimisha","year":"0","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref220","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.355"},{"key":"ref222","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00660"},{"key":"ref221","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-11021-5_17"},{"key":"ref229","first-page":"2063","article-title":"Transfer learning for related reinforcement learning tasks via image-to-image translation","author":"gamrian","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref228","first-page":"747","article-title":"Improving surgical training phantoms by hyperrealism: Deep unpaired image-to-image translation from real surgeries","author":"engelhardt","year":"0","journal-title":"Proc Int Conf Med Image Comput Comput - Assist Interv"},{"key":"ref227","doi-asserted-by":"publisher","DOI":"10.1007\/s12194-019-00520-y"},{"key":"ref226","doi-asserted-by":"publisher","DOI":"10.23919\/EUSIPCO.2019.8902799"},{"key":"ref225","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00028"},{"key":"ref224","article-title":"Harmonic unpaired image-to-image translation","author":"zhang","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref223","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.3301825"},{"key":"ref127","first-page":"4720","article-title":"Reversible GANs for memory-efficient image-to-image translation","author":"ouderaa","year":"0","journal-title":"Proc IEEE Conf Comput Vis Pattern Recognit"},{"key":"ref126","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01089"},{"key":"ref125","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00484"},{"key":"ref124","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00380"},{"key":"ref129","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00532"},{"key":"ref128","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.5765"},{"key":"ref130","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00819"},{"key":"ref133","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00152"},{"key":"ref134","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00155"},{"key":"ref131","first-page":"195","article-title":"Augmented cyclegan: Learning many-to-many mappings from unpaired data","volume":"80","author":"almahairi","year":"0","journal-title":"Mach Learn Res"},{"key":"ref132","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00579"},{"key":"ref232","first-page":"172","article-title":"Generalizing a person retrieval model hetero- and homogeneously","author":"zhong","year":"0","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref233","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00224"},{"key":"ref230","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00110"},{"key":"ref231","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00541"},{"key":"ref239","doi-asserted-by":"publisher","DOI":"10.1145\/3394171.3416270"},{"key":"ref238","doi-asserted-by":"publisher","DOI":"10.1145\/3240508.3240704"},{"key":"ref235","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.175"},{"key":"ref234","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00737"},{"key":"ref237","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR48806.2021.9412890"},{"key":"ref236","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00955"},{"key":"ref136","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01231-1_14"},{"key":"ref135","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58598-3_34"},{"key":"ref138","first-page":"15885","article-title":"Few-shot image generation with elastic weight consolidation","author":"li","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref137","first-page":"11507","article-title":"Learning to transfer: Unsupervised meta domain translation","volume":"34","author":"lin","year":"2020","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"ref139","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01060"},{"key":"ref140","first-page":"2104","article-title":"One-shot unsupervised cross domain translation","author":"benaim","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref141","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00187"},{"key":"ref142","doi-asserted-by":"publisher","DOI":"10.1145\/1459352.1459355"},{"key":"ref143","first-page":"1","article-title":"Small data challenges in big data era: A survey of recent progress on unsupervised and semi-supervised methods","author":"qi","year":"2020","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"ref144","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3084358"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref145","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btr502"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref241","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00870"},{"key":"ref242","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00245"},{"key":"ref243","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00600"},{"key":"ref244","first-page":"717","article-title":"XingGAN for person image generation","author":"tang","year":"0","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref240","article-title":"Edge guided GANs with semantic preserving for semantic image synthesis","author":"tang","year":"2020"},{"key":"ref248","article-title":"DF-GAN: Deep fusion generative adversarial networks for text-to-image synthesis","author":"tao","year":"2020"},{"key":"ref247","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2019.8756586"},{"key":"ref246","doi-asserted-by":"publisher","DOI":"10.1145\/3343031.3350980"},{"key":"ref245","first-page":"0","article-title":"Bipartite graph reasoning GANs for person image generation","author":"tang","year":"0","journal-title":"Proc BMVC"},{"key":"ref249","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00790"},{"key":"ref109","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01240-3_12"},{"key":"ref108","article-title":"Unsupervised cross-domain image generation","author":"taigman","year":"0","journal-title":"Proc 5th Int Conf Learn Representations"},{"key":"ref107","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58523-5_35"},{"key":"ref106","first-page":"1287","article-title":"Image-to-image translation for cross-domain disentanglement","author":"gonzalez-garcia","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref105","article-title":"PixeINN: Example-based image synthesis","author":"bansal","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref104","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01464"},{"key":"ref103","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01130"},{"key":"ref102","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00519"},{"key":"ref111","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00919"},{"key":"ref112","first-page":"800","article-title":"Unpaired image-to-image translation using adversarial consistency loss","author":"zhao","year":"0","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref110","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01258-8_40"},{"key":"ref250","first-page":"5967","article-title":"Fader networks: Manipulating images by sliding attributes","author":"lample","year":"0","journal-title":"Proc Adv Neural Inf Process Syst 30"},{"key":"ref251","article-title":"Emerging disentanglement in auto-encoder based unsupervised image content transfer","author":"press","year":"0","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref254","doi-asserted-by":"publisher","DOI":"10.1145\/3240508.3240612"},{"key":"ref255","doi-asserted-by":"publisher","DOI":"10.1145\/3394171.3413981"},{"key":"ref252","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2019.8803654"},{"key":"ref253","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00740"},{"key":"ref257","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2019.2895292"},{"key":"ref256","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2020.103886"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107343"},{"key":"ref258","article-title":"Adversarial shape learning for building extraction in VHR remote sensing images","author":"ding","year":"2021"},{"key":"ref11","first-page":"1857","article-title":"Learning to discover cross-domain relations with generative adversarial networks","author":"kim","year":"0","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.310"},{"key":"ref13","first-page":"3693","article-title":"Unsupervised attention-guided image-to-image translation","author":"mejjati","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00600"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.278"},{"key":"ref118","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58580-8_13"},{"key":"ref16","first-page":"465","article-title":"Toward multimodal image-to-image translation","author":"zhu","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref117","first-page":"7198","article-title":"Swapping autoencoder for deep image manipulation","volume":"33","author":"park","year":"2020","journal-title":"Adv Neural Inf Proc Syst"},{"key":"ref17","first-page":"3","article-title":"Contextual-based image inpainting: Infer, match, and translate","author":"song","year":"0","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.01065"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00578"},{"key":"ref119","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01614"},{"key":"ref114","article-title":"One-sided unsupervised domain mapping","volume":"30","author":"benaim","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref113","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58536-5_40"},{"key":"ref116","first-page":"319","article-title":"Contrastive learning for unpaired image-to-image translation","author":"park","year":"0","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref115","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00253"},{"key":"ref120","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00927"},{"key":"ref121","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00593"},{"key":"ref122","first-page":"164","article-title":"Attention-GAN for object transfiguration in wild images","author":"chen","year":"0","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref123","article-title":"InstaGAN: Instance-aware image-to-image translation","author":"mo","year":"0","journal-title":"Proc Int Conf Learn Representations"}],"container-title":["IEEE Transactions on Multimedia"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6046\/9687854\/09528943.pdf?arnumber=9528943","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,21]],"date-time":"2024-02-21T18:56:02Z","timestamp":1708541762000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9528943\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"references-count":258,"URL":"https:\/\/doi.org\/10.1109\/tmm.2021.3109419","relation":{},"ISSN":["1520-9210","1941-0077"],"issn-type":[{"value":"1520-9210","type":"print"},{"value":"1941-0077","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]}}}