{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,17]],"date-time":"2026-03-17T17:13:06Z","timestamp":1773767586738,"version":"3.50.1"},"reference-count":60,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61672202"],"award-info":[{"award-number":["61672202"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"State Key Program of NSFC-Shenzhen Joint Foundation","doi-asserted-by":"publisher","award":["U1613217"],"award-info":[{"award-number":["U1613217"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities of China","doi-asserted-by":"publisher","award":["PA2019GDPK0076"],"award-info":[{"award-number":["PA2019GDPK0076"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2020]]},"DOI":"10.1109\/access.2020.3021531","type":"journal-article","created":{"date-parts":[[2020,9,3]],"date-time":"2020-09-03T22:23:00Z","timestamp":1599171780000},"page":"161509-161518","source":"Crossref","is-referenced-by-count":21,"title":["LAUN Improved StarGAN for Facial Emotion Recognition"],"prefix":"10.1109","volume":"8","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1751-2291","authenticated-orcid":false,"given":"Xiaohua","family":"Wang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4362-5804","authenticated-orcid":false,"given":"Jianqiao","family":"Gong","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2122-0240","authenticated-orcid":false,"given":"Min","family":"Hu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6939-0850","authenticated-orcid":false,"given":"Yu","family":"Gu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4860-9184","authenticated-orcid":false,"given":"Fuji","family":"Ren","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1145\/3343031.3351032"},{"key":"ref38","first-page":"1","article-title":"ExprGAN: Facial expression editing with controllable expression intensity","author":"ding","year":"2018","journal-title":"Proc 32nd AAAI Conf Artif Intell"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1145\/2818346.2830595"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2938194"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2980024"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2985726"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.463"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00354"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00693"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00286"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.243"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/PIC.2016.7949500"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/TCYB.2013.2247592"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.5120\/18171-9055"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.7551\/mitpress\/1140.001.0001"},{"key":"ref1","first-page":"193","article-title":"Communication without words","volume":"6","author":"mehrabian","year":"2008","journal-title":"Commun Theory"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2710635"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/MIPR.2019.00025"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ICSGEA.2019.00052"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.1994.576366"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1049\/ji-3-2.1946.0074"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2008.52"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2005.177"},{"key":"ref50","first-page":"1790","article-title":"Deep convolutional neural network for image deconvolution","author":"xu","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00813"},{"key":"ref59","article-title":"Inception-v4, inception-ResNet and the impact of residual connections on learning","author":"szegedy","year":"2016","journal-title":"arXiv 1602 07261"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2017.05.025"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2017.7966040"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1017\/S1355617717000145"},{"key":"ref55","first-page":"6626","article-title":"Gans trained by a two time-scale update rule converge to a local Nash equilibrium","author":"heusel","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2005.1521424"},{"key":"ref53","article-title":"The Karolinska directed emotional faces&#x2014;KDEF","author":"lundqvist","year":"1998"},{"key":"ref52","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"2014","journal-title":"arXiv 1412 6980"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICDAR.2003.1227801"},{"key":"ref11","article-title":"Facial affect detection using convolutional neural networks","author":"wang","year":"2016"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2019.107111"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2936976"},{"key":"ref13","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.141"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00916"},{"key":"ref17","first-page":"681","article-title":"Robust real-time face pose and facial expression recovery","volume":"1","author":"zhu","year":"2006","journal-title":"Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit (CVPR)"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/2845089"},{"key":"ref19","article-title":"Very deep convolutional networks for large-scale image recognition","author":"simonyan","year":"2014","journal-title":"arXiv 1409 1556"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1360\/N972017-01272"},{"key":"ref3","author":"ekman","year":"1997","journal-title":"What the Face Reveals Basic and Applied Studies of Spontaneous Expression using the Facial Action Coding System (FACS)"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46487-9_19"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2014.2316151"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/MVA.2015.7153185"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ISBI.2018.8363678"},{"key":"ref49","article-title":"Neural machine translation by jointly learning to align and translate","author":"bahdanau","year":"2014","journal-title":"arXiv 1409 0473"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/IC3I.2016.7918019"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01264-9_47"},{"key":"ref45","first-page":"10541","article-title":"Large scale adversarial representation learning","author":"donahue","year":"2019","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref48","first-page":"234","article-title":"U-Net: Convolutional networks for biomedical image segmentation","author":"ronneberger","year":"2015","journal-title":"Proc Int Conf Med Image Comput Comput -Assist Intervent"},{"key":"ref47","article-title":"Attention U-Net: Learning where to look for the pancreas","author":"oktay","year":"2018","journal-title":"arXiv 1804 03999"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01219-9_11"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00453"},{"key":"ref43","article-title":"Unsupervised representation learning with deep convolutional generative adversarial networks","author":"radford","year":"2015","journal-title":"arXiv 1511 06434"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/8948470\/09186113.pdf?arnumber=9186113","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,12,17]],"date-time":"2021-12-17T20:02:52Z","timestamp":1639771372000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9186113\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"references-count":60,"URL":"https:\/\/doi.org\/10.1109\/access.2020.3021531","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]}}}