{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T00:40:13Z","timestamp":1774917613007,"version":"3.50.1"},"reference-count":65,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"1","license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Fundamental Research Grant Scheme (FRGS) MoHE from the Ministry of Education Malaysia","award":["FP004-2016"],"award-info":[{"award-number":["FP004-2016"]}]},{"name":"UM Frontier Research from University of Malaya","award":["FG002-17AFR"],"award-info":[{"award-number":["FG002-17AFR"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. on Image Process."],"published-print":{"date-parts":[[2019,1]]},"DOI":"10.1109\/tip.2018.2866698","type":"journal-article","created":{"date-parts":[[2018,8,22]],"date-time":"2018-08-22T18:47:12Z","timestamp":1534963632000},"page":"394-409","source":"Crossref","is-referenced-by-count":90,"title":["Improved ArtGAN for Conditional Synthesis of Natural Image and Artwork"],"prefix":"10.1109","volume":"28","author":[{"given":"Wei Ren","family":"Tan","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7677-2865","authenticated-orcid":false,"given":"Chee Seng","family":"Chan","sequence":"additional","affiliation":[]},{"given":"Hernan E.","family":"Aguirre","sequence":"additional","affiliation":[]},{"given":"Kiyoshi","family":"Tanaka","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","first-page":"1","article-title":"Improving generative adversarial networks with denoising feature matching","author":"warde-farley","year":"2017","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.304"},{"key":"ref33","first-page":"4790","article-title":"Conditional image generation with PixelCNN decoders","author":"van den oord","year":"2016","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref32","first-page":"1747","article-title":"Pixel recurrent neural networks","author":"van den oord","year":"2016","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref31","first-page":"1462","article-title":"DRAW: A recurrent neural network for image generation","author":"gregor","year":"2015","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref30","first-page":"899","article-title":"Generalized denoising auto-encoders as generative models","author":"bengio","year":"2013","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref37","author":"qi","year":"2017","journal-title":"Loss-sensitive generative adversarial networks on lipschitz densities"},{"key":"ref36","first-page":"5767","article-title":"Improved training of Wasserstein GANs","author":"gulrajani","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref35","first-page":"1","article-title":"Amortised map inference for image super-resolution","author":"s\u00f8nderby","year":"2017","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref34","author":"salimans","year":"2017","journal-title":"Pixelcnn++ Improving the pixelcnn with discretized logistic mixture likelihood and other modifications"},{"key":"ref60","first-page":"1","article-title":"Adversarially learned inference","author":"dumoulin","year":"2017","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref62","first-page":"1","article-title":"Calibrating energy-based generative adversarial networks","author":"dai","year":"2017","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref61","first-page":"1","article-title":"Generative multi-adversarial networks","author":"durugkar","year":"2017","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref63","first-page":"1","article-title":"LR-GAN: Layered recursive generative adversarial networks for image generation","author":"jianwei","year":"2017","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref28","author":"kingma","year":"2013","journal-title":"Auto-encoding variational bayes"},{"key":"ref64","author":"wang","year":"2016","journal-title":"Learning to draw samples With application to amortized mle for generative adversarial learning"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2012.2192125"},{"key":"ref65","author":"eliot","year":"1921","journal-title":"The Sacred Wood Essays on Poetry and Criticism"},{"key":"ref29","first-page":"3581","article-title":"Semi-supervised learning with deep generative models","author":"kingma","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref2","author":"simonyan","year":"2014","journal-title":"Very Deep Convolutional Networks for Large-scale Image Recognition"},{"key":"ref1","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICVGIP.2008.47"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2017.8296985"},{"key":"ref21","article-title":"The Caltech-UCSD birds-200-2011 dataset","author":"wah","year":"2011"},{"key":"ref24","first-page":"2226","article-title":"Improved techniques for training gans","author":"salimans","year":"2016","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref23","author":"saleh","year":"2015","journal-title":"Large-scale classification of fine-art paintings Learning the right metric on the right feature"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1162\/NECO_a_00086"},{"key":"ref25","first-page":"2002","article-title":"Generating more realistic images using gated MRF&#x2019;s","author":"mnih","year":"2010","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.374"},{"key":"ref51","author":"kingma","year":"2014","journal-title":"Adam A method for stochastic optimization"},{"key":"ref59","first-page":"1","article-title":"Learning to generate samples from noise through infusion training","author":"bordes","year":"2017","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1214\/aoms\/1177704472"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.202"},{"key":"ref56","first-page":"1","volume":"1","author":"odena","year":"2016","journal-title":"Deconvolution and checkerboard artifacts"},{"key":"ref55","first-page":"3703","article-title":"Ceci n&#x2019;est pas une pipe: A deep convolutional network for fine-art paintings classification","author":"tan","year":"2016","journal-title":"Proc IEEE Int Conf Image Process"},{"key":"ref54","author":"theis","year":"2015","journal-title":"A note on the evaluation of generative models"},{"key":"ref53","author":"abadi","year":"2016","journal-title":"Tensorflow Large-scale machine learning on heterogeneous distributed systems"},{"key":"ref52","author":"goodfellow","year":"2016","journal-title":"Nips 2016 tutorial Generative adversarial networks"},{"key":"ref10","first-page":"1060","article-title":"Generative adversarial text to image synthesis","author":"reed","year":"2016","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.629"},{"key":"ref40","first-page":"1","article-title":"Progressive growing of GANs for improved quality, stability, and variation","author":"karras","year":"2018","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref12","article-title":"The mnist database of handwritten digits","author":"lecun","year":"1998"},{"key":"ref13","article-title":"Learning multiple layers of features from tiny images","author":"krizhevsky","year":"2009"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref15","first-page":"1","article-title":"Energy-based generative adversarial network","author":"zhao","year":"2017","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref16","author":"berthelot","year":"2017","journal-title":"BEGAN Boundary Equilibrium Generative Adversarial Networks"},{"key":"ref17","author":"wang","year":"2016","journal-title":"Ensembles of generative adversarial networks"},{"key":"ref18","first-page":"1","article-title":"Snapshot ensembles: Train 1, get M for free","author":"huang","year":"2017","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref19","first-page":"215","article-title":"An analysis of single-layer networks in unsupervised feature learning","author":"coates","year":"2011","journal-title":"Proc 14th Int Conf Artif Intell Statist"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref6","author":"radford","year":"2015","journal-title":"Unsupervised Representation learning with deep convolutional generative adversarial networks CoRR"},{"key":"ref5","author":"mirza","year":"2014","journal-title":"Conditional generative adversarial nets"},{"key":"ref8","first-page":"1486","article-title":"Deep generative image models using a Laplacian pyramid of adversarial networks","author":"denton","year":"2015","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref7","author":"springenberg","year":"2015","journal-title":"Unsupervised and semi-supervised learning with categorical generative adversarial networks"},{"key":"ref49","first-page":"3387","article-title":"Synthesizing the preferred inputs for neurons in neural networks via deep generator networks","author":"nguyen","year":"2016","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref9","first-page":"2642","article-title":"Conditional image synthesis with auxiliary classifier GANs","author":"odena","year":"2017","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref46","first-page":"2172","article-title":"InfoGAN: Interpretable representation learning by information maximizing generative adversarial nets","author":"chen","year":"2016","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref45","first-page":"1","article-title":"Conditional generative adversarial nets for convolutional face generation","volume":"2015","author":"gauthier","year":"2015","journal-title":"Class Project Stanford CS231N Convolutional Neural Netw Vis Recognit Winter Semester"},{"key":"ref48","article-title":"Labeled faces in the wild: Updates and new reporting procedures","author":"huang","year":"2014"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.299"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.278"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2678168"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2651375"}],"container-title":["IEEE Transactions on Image Processing"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/83\/8468142\/08444471.pdf?arnumber=8444471","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,13]],"date-time":"2022-07-13T21:08:02Z","timestamp":1657746482000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8444471\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,1]]},"references-count":65,"journal-issue":{"issue":"1"},"URL":"https:\/\/doi.org\/10.1109\/tip.2018.2866698","relation":{},"ISSN":["1057-7149","1941-0042"],"issn-type":[{"value":"1057-7149","type":"print"},{"value":"1941-0042","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019,1]]}}}