{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T03:10:05Z","timestamp":1774926605534,"version":"3.50.1"},"reference-count":115,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"7","license":[{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,7,1]],"date-time":"2024-07-01T00:00:00Z","timestamp":1719792000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Pioneer and Leading Goose Research and Development Program of Zhejiang","award":["2023C01212"],"award-info":[{"award-number":["2023C01212"]}]},{"name":"Young Elite Scientists Sponsorship Program by China Association for Science and Technology","award":["2022QNRC001"],"award-info":[{"award-number":["2022QNRC001"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61976188"],"award-info":[{"award-number":["61976188"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Circuits Syst. Video Technol."],"published-print":{"date-parts":[[2024,7]]},"DOI":"10.1109\/tcsvt.2023.3348471","type":"journal-article","created":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T20:02:10Z","timestamp":1704139330000},"page":"6167-6180","source":"Crossref","is-referenced-by-count":6,"title":["Statistics Enhancement Generative Adversarial Networks for Diverse Conditional Image Synthesis"],"prefix":"10.1109","volume":"34","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0062-6308","authenticated-orcid":false,"given":"Zhiwen","family":"Zuo","sequence":"first","affiliation":[{"name":"College of Computer Science and Technology, Zhejiang Gongshang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4811-6320","authenticated-orcid":false,"given":"Ailin","family":"Li","sequence":"additional","affiliation":[{"name":"College of Computer Science and Technology, Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4484-5112","authenticated-orcid":false,"given":"Zhizhong","family":"Wang","sequence":"additional","affiliation":[{"name":"College of Computer Science and Technology, Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4791-454X","authenticated-orcid":false,"given":"Lei","family":"Zhao","sequence":"additional","affiliation":[{"name":"College of Computer Science and Technology, Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5244-3274","authenticated-orcid":false,"given":"Jianfeng","family":"Dong","sequence":"additional","affiliation":[{"name":"College of Computer Science and Technology, Zhejiang Gongshang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5566-4689","authenticated-orcid":false,"given":"Xun","family":"Wang","sequence":"additional","affiliation":[{"name":"College of Computer Science and Technology, Zhejiang Gongshang University, Hangzhou, China"}]},{"given":"Meng","family":"Wang","sequence":"additional","affiliation":[{"name":"School of Computer Science and Information Engineering, Hefei University of Technology, Hefei, China"}]}],"member":"263","reference":[{"key":"ref1","first-page":"1106","article-title":"ImageNet classification with deep convolutional neural networks","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"25","author":"Krizhevsky"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.81"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.91"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref5","article-title":"An image is worth 16\u00d716 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020","journal-title":"arXiv:2010.11929"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6829"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01348"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01326"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19830-4_29"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/tcsvt.2023.3283430"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.374"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00877"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00175"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2022.3150959"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3059295"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.5555\/2969033.2969125"},{"key":"ref18","article-title":"Conditional generative adversarial nets","author":"Mirza","year":"2014","journal-title":"arXiv:1411.1784"},{"key":"ref19","first-page":"2642","article-title":"Conditional image synthesis with auxiliary classifier GANs","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Odena"},{"key":"ref20","article-title":"Large scale GAN training for high fidelity natural image synthesis","volume-title":"arXiv:1809.11096","author":"Brock","year":"2018"},{"key":"ref21","article-title":"CGANs with projection discriminator","author":"Miyato","year":"2018","journal-title":"arXiv:1802.05637"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2020.3037688"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-019-01284-z"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58545-7_19"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19787-1_25"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.278"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00457"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2020.3001267"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/WACV51458.2022.00323"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.19"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1809.00219"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2022.3188433"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2022.3192099"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00976"},{"key":"ref37","article-title":"Deep multi-scale video prediction beyond mean square error","author":"Mathieu","year":"2015","journal-title":"arXiv:1511.05440"},{"key":"ref38","first-page":"465","article-title":"Toward multimodal image-to-image translation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Zhu"},{"key":"ref39","article-title":"NIPS 2016 tutorial: Generative adversarial networks","author":"Goodfellow","year":"2017","journal-title":"arXiv:1701.00160"},{"key":"ref40","article-title":"Unrolled generative adversarial networks","author":"Metz","year":"2016","journal-title":"arXiv:1611.02163"},{"key":"ref41","first-page":"2234","article-title":"Improved techniques for training GANs","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Salimans"},{"key":"ref42","first-page":"3308","article-title":"VEEGAN: Reducing mode collapse in GANs using implicit variational learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Srivastava"},{"key":"ref43","first-page":"2275","article-title":"BourGAN: Generative networks with metric embeddings","volume-title":"Proc. 32nd Int. Conf. Neural Inf. Process. Syst.","author":"Xiao"},{"key":"ref44","first-page":"1498","article-title":"PacGAN: The power of two samples in generative adversarial networks","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Lin"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01219-9_11"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01246-5_3"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00578"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01461"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1312.6114"},{"key":"ref51","first-page":"1278","article-title":"Stochastic backpropagation and approximate inference in deep generative models","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Rezende"},{"key":"ref52","first-page":"1558","article-title":"Autoencoding beyond pixels using a learned similarity metric","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Larsen"},{"key":"ref53","article-title":"Adversarially learned inference","author":"Dumoulin","year":"2016","journal-title":"arXiv:1606.00704"},{"key":"ref54","article-title":"Adversarial feature learning","author":"Donahue","year":"2016","journal-title":"arXiv:1605.09782"},{"key":"ref55","first-page":"2172","article-title":"InfoGAN: Interpretable representation learning by information maximizing generative adversarial nets","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Chen"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00152"},{"key":"ref57","article-title":"Diversity-sensitive conditional generative adversarial networks","author":"Yang","year":"2019","journal-title":"arXiv:1901.09024"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01611"},{"key":"ref59","article-title":"Large scale image completion via co-modulated generative adversarial networks","volume-title":"Proc. Int. Conf. Learn. Represent.","author":"Zhao"},{"key":"ref60","article-title":"MINE: Mutual information neural estimation","author":"Belghazi","year":"2018","journal-title":"arXiv:1801.04062"},{"key":"ref61","article-title":"Representation learning with contrastive predictive coding","author":"van den Oord","year":"2018","journal-title":"arXiv:1807.03748"},{"key":"ref62","article-title":"Learning deep representations by mutual information estimation and maximization","author":"Hjelm","year":"2018","journal-title":"arXiv:1808.06670"},{"key":"ref63","article-title":"Unsupervised representation learning with deep convolutional generative adversarial networks","author":"Radford","year":"2015","journal-title":"arXiv:1511.06434"},{"key":"ref64","article-title":"Deep generative image models using a Laplacian pyramid of adversarial networks","author":"Denton","year":"2015","journal-title":"arXiv:1506.05751"},{"key":"ref65","article-title":"Progressive growing of GANs for improved quality, stability, and variation","author":"Karras","year":"2017","journal-title":"arXiv:1710.10196"},{"key":"ref66","first-page":"7354","article-title":"Self-attention generative adversarial networks","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Zhang"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00453"},{"key":"ref68","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.2106.12423"},{"key":"ref70","first-page":"14745","article-title":"TransGan: Two pure transformers can make one strong GAN, and that can scale up","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Jiang"},{"key":"ref71","doi-asserted-by":"publisher","DOI":"10.1145\/3528233.3530738"},{"key":"ref72","article-title":"Geometric GAN","author":"Lim","year":"2017","journal-title":"arXiv:1705.02894"},{"key":"ref73","article-title":"Wasserstein GAN","author":"Arjovsky","year":"2017","journal-title":"arXiv:1701.07875"},{"key":"ref74","first-page":"5767","article-title":"Improved training of Wasserstein GANs","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Gulrajani"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.304"},{"key":"ref76","first-page":"3481","article-title":"Which training methods for GANs do actually converge?","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Mescheder"},{"key":"ref77","article-title":"Spectral normalization for generative adversarial networks","author":"Miyato","year":"2018","journal-title":"arXiv:1802.05957"},{"key":"ref78","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01243"},{"key":"ref79","article-title":"Consistency regularization for generative adversarial networks","author":"Zhang","year":"2019","journal-title":"arXiv:1910.12027"},{"key":"ref80","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00648"},{"key":"ref81","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN48605.2020.9207181"},{"key":"ref82","article-title":"GANs with variational entropy regularizers: Applications in mitigating the mode-collapse issue","author":"Khorramshahi","year":"2020","journal-title":"arXiv:2009.11921"},{"key":"ref83","article-title":"Combating mode collapse in GAN training: An empirical analysis using Hessian eigenvalues","author":"Durall","year":"2020","journal-title":"arXiv:2012.09673"},{"key":"ref84","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00666"},{"key":"ref85","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2020.3045000"},{"key":"ref86","first-page":"37690","article-title":"UniGAN: Reducing mode collapse in GANs using a uniform generator","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Pan"},{"key":"ref87","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2023.103863"},{"key":"ref88","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58574-7_33"},{"key":"ref89","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2021.12.095"},{"key":"ref90","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00888"},{"key":"ref91","article-title":"MGAN: Training generative adversarial nets with multiple generators","volume-title":"Proc. Int. Conf. Learn. Represent.","author":"Hoang"},{"key":"ref92","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107646"},{"key":"ref93","first-page":"224","article-title":"Generalization and equilibrium in generative adversarial nets (GANs)","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Arora"},{"key":"ref94","article-title":"On the limitations of first-order approximation in GAN dynamics","author":"Li","year":"2017","journal-title":"arXiv:1706.09884"},{"key":"ref95","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN52387.2021.9533656"},{"key":"ref96","article-title":"Overcoming mode collapse with adaptive multi adversarial training","author":"Mangalam","year":"2021","journal-title":"arXiv:2112.14406"},{"key":"ref97","doi-asserted-by":"publisher","DOI":"10.1109\/tcsvt.2023.3311039"},{"key":"ref98","doi-asserted-by":"publisher","DOI":"10.1109\/tcsvt.2023.3298811"},{"key":"ref99","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2022.3170588"},{"key":"ref100","doi-asserted-by":"publisher","DOI":"10.1109\/WACV48630.2021.00399"},{"key":"ref101","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2022.3192907"},{"key":"ref102","doi-asserted-by":"publisher","DOI":"10.1002\/cpa.3160360204"},{"key":"ref103","first-page":"271","article-title":"f-GAN: Training generative neural samplers using variational divergence minimization","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Nowozin"},{"key":"ref104","article-title":"Learning multiple layers of features from tiny images","author":"Krizhevsky","year":"2009"},{"key":"ref105","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00143"},{"key":"ref106","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-40602-7_39"},{"key":"ref107","article-title":"Caltech-UCSD birds 200","author":"Welinder","year":"2010"},{"key":"ref108","first-page":"6626","article-title":"GANs trained by a two time-scale update rule converge to a local nash equilibrium","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Heusel"},{"key":"ref109","article-title":"Assessing generative models via precision and recall","volume":"31","author":"Sajjadi","year":"2018","journal-title":"Proc. Adv. Neural Inf. Process. Syst."},{"key":"ref110","first-page":"3927","article-title":"Improved precision and recall metric for assessing generative models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"32","author":"Kynk\u00e4\u00e4nniemi"},{"key":"ref111","first-page":"5847","article-title":"On GANs and GMMs","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Richardson"},{"key":"ref112","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"ref113","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref114","first-page":"698","article-title":"Are GANs created equal? A large-scale study","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"31","author":"Lucic"},{"key":"ref115","first-page":"5171","article-title":"On variational bounds of mutual information","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Poole"}],"container-title":["IEEE Transactions on Circuits and Systems for Video Technology"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/76\/10582818\/10378665.pdf?arnumber=10378665","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,5]],"date-time":"2024-09-05T18:48:03Z","timestamp":1725562083000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10378665\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7]]},"references-count":115,"journal-issue":{"issue":"7"},"URL":"https:\/\/doi.org\/10.1109\/tcsvt.2023.3348471","relation":{},"ISSN":["1051-8215","1558-2205"],"issn-type":[{"value":"1051-8215","type":"print"},{"value":"1558-2205","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7]]}}}