{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,30]],"date-time":"2026-04-30T23:58:48Z","timestamp":1777593528917,"version":"3.51.4"},"reference-count":53,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"name":"Symbiosis International (Deemed University), Pune, India"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2021]]},"DOI":"10.1109\/access.2021.3112996","type":"journal-article","created":{"date-parts":[[2021,9,15]],"date-time":"2021-09-15T20:03:53Z","timestamp":1631736233000},"page":"131583-131613","source":"Crossref","is-referenced-by-count":54,"title":["Neural Style Transfer: A Critical Review"],"prefix":"10.1109","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8959-8659","authenticated-orcid":false,"given":"Akhil","family":"Singh","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7352-5236","authenticated-orcid":false,"given":"Vaibhav","family":"Jaiswal","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4404-2985","authenticated-orcid":false,"given":"Gaurav","family":"Joshi","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8092-5625","authenticated-orcid":false,"given":"Adith","family":"Sanjeeve","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3882-7030","authenticated-orcid":false,"given":"Shilpa","family":"Gite","sequence":"additional","affiliation":[]},{"given":"Ketan","family":"Kotecha","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","article-title":"GANcraft: Unsupervised 3D neural rendering of minecraft worlds","author":"hao","year":"2021","journal-title":"arXiv 2104 07659"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00696"},{"key":"ref33","article-title":"Kunster&#x2014;AR art video maker&#x2014;Real time video neural style transfer on mobile devices","author":"dudzik","year":"2020","journal-title":"arXiv 2005 03415"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.745"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-45886-1_3"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.18488\/journal.76.2020.72.86.95"},{"key":"ref37","article-title":"Bringing impressionism to life with neural style transfer in come swim","author":"joshi","year":"2017","journal-title":"arXiv 1701 04928"},{"key":"ref36","article-title":"An empirical study on evaluation metrics of generative adversarial networks","author":"xu","year":"2018","journal-title":"arXiv 1806 07755"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.126"},{"key":"ref34","article-title":"Multi-style generative network for real-time transfer","author":"zhang","year":"2017","journal-title":"arXiv 1703 06953"},{"key":"ref28","article-title":"Exploring style transfer: Extensions to neural style transfer","author":"makow","year":"2017"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1145\/3306305.3332370"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-16-0708-0_2"},{"key":"ref2","year":"2021","journal-title":"Adobe Premiere Pro"},{"key":"ref1","year":"2021","journal-title":"The Smartphone vs the Camera Industry"},{"key":"ref20","article-title":"Depth-aware arbitrary style transfer using instance normalization","author":"kitov","year":"2019","journal-title":"arXiv 1906 01123"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/I-SPAN.2018.00039"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.296"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.438"},{"key":"ref23","article-title":"Controlling perceptual factors in neural style transfer","author":"gatys","year":"2016","journal-title":"arXiv 1611 07865"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.740"},{"key":"ref25","article-title":"Preserving color in neural artistic style transfer","author":"gatys","year":"2016","journal-title":"arXiv 1606 05897"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.3390\/su11205673"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.2970919"},{"key":"ref53","article-title":"Visualizing and understanding generative adversarial networks extended abstract","author":"bau","year":"2019","journal-title":"arXiv 1901 09887"},{"key":"ref52","article-title":"ReCoNet: Real-time coherent video style transfer network","author":"gao","year":"2018","journal-title":"arXiv 1807 01197"},{"key":"ref10","article-title":"Discriminative unsupervised feature learning with exemplar convolutional neural networks","author":"dosovitskiy","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst (NIPS)"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref40","article-title":"PSGAN: Pose and expression robust spatial-aware GAN for customizable makeup transfer","author":"jiang","year":"2019","journal-title":"arXiv 1909 06956"},{"key":"ref12","article-title":"Image-to-image translation with conditional adversarial networks","author":"isola","year":"2016","journal-title":"ArXiv 1611 07004"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.265"},{"key":"ref14","article-title":"Towards the automatic anime characters creation with generative adversarial networks","author":"jin","year":"2017","journal-title":"arXiv 1708 05509"},{"key":"ref15","article-title":"On convergence and stability of GANs","author":"kodali","year":"2017","journal-title":"arXiv 1705 07215"},{"key":"ref16","article-title":"Photo-realistic single image super-resolution using a generative adversarial network","author":"ledig","year":"2016","journal-title":"arXiv 1609 04802"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00986"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2018.8546172"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46487-9_43"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2019.2921336"},{"key":"ref3","year":"2021","journal-title":"DaVinci Resolve"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1088\/1742-6596\/1651\/1\/012156"},{"key":"ref5","author":"li","year":"2019","journal-title":"A Literature Review of Neural Style Transfer"},{"key":"ref8","article-title":"A style-based generator architecture for generative adversarial networks","author":"karras","year":"2018","journal-title":"arXiv 1812 04948"},{"key":"ref7","article-title":"Generative adversarial networks","author":"goodfellow","year":"2014","journal-title":"arXiv 1406 2661"},{"key":"ref49","year":"2016","journal-title":"Computer Vision&#x2014;ECCV 2016"},{"key":"ref9","article-title":"Unsupervised representation learning with deep convolutional generative adversarial networks","author":"radford","year":"2015","journal-title":"arXiv 1511 06434"},{"key":"ref46","article-title":"StyTr&#x00B2;: Unbiased image style transfer with transformers","author":"deng","year":"2021","journal-title":"arXiv 2105 14576"},{"key":"ref45","article-title":"Unpaired high-resolution and scalable style transfer using generative adversarial networks","author":"junginger","year":"2018","journal-title":"arXiv 1810 05724"},{"key":"ref48","year":"2019","journal-title":"Computer Vision&#x2014;ECCV 2018 Workshops"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.397"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1145\/3386569.3392453"},{"key":"ref41","article-title":"Lipstick ain&#x2019;t enough: Beyond color matching for in-the-wild makeup transfer","author":"nguyen","year":"2021","journal-title":"arXiv 2104 01867"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1145\/3397166.3409140"},{"key":"ref43","article-title":"Federated CycleGAN for privacy-preserving image-to-image translation","author":"song","year":"2021","journal-title":"arXiv 2106 09246"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/9312710\/09539183.pdf?arnumber=9539183","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,12,17]],"date-time":"2021-12-17T19:55:37Z","timestamp":1639770937000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9539183\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"references-count":53,"URL":"https:\/\/doi.org\/10.1109\/access.2021.3112996","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]}}}