{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,4]],"date-time":"2025-06-04T18:10:08Z","timestamp":1749060608172,"version":"3.41.0"},"reference-count":55,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"6","license":[{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U22A2096","62036007"],"award-info":[{"award-number":["U22A2096","62036007"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shaanxi Province Core Technology Research and Development Project","award":["2024QY2-GJHX-11"],"award-info":[{"award-number":["2024QY2-GJHX-11"]}]},{"DOI":"10.13039\/100017367","name":"Young Talent Fund of Association for Science and Technology, Shaanxi, China","doi-asserted-by":"publisher","award":["20230121"],"award-info":[{"award-number":["20230121"]}],"id":[{"id":"10.13039\/100017367","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Chinese Computer Federation (CCF)-Baidu Open Fund"},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["QTZX23042"],"award-info":[{"award-number":["QTZX23042"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Neural Netw. Learning Syst."],"published-print":{"date-parts":[[2025,6]]},"DOI":"10.1109\/tnnls.2025.3559477","type":"journal-article","created":{"date-parts":[[2025,4,28]],"date-time":"2025-04-28T17:34:08Z","timestamp":1745861648000},"page":"10015-10027","source":"Crossref","is-referenced-by-count":0,"title":["PStyle-3D: Example-Based 3-D-Aware Portrait Style Domain Adaptation"],"prefix":"10.1109","volume":"36","author":[{"given":"Chaohua","family":"Shi","sequence":"first","affiliation":[{"name":"State Key Laboratory of Integrated Services Networks (ISN), School of Telecommunications Engineering, Xidian University, Xi&#x2019;an, Shaanxi, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4179-7701","authenticated-orcid":false,"given":"Mingrui","family":"Zhu","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Integrated Services Networks (ISN), School of Telecommunications Engineering, Xidian University, Xi&#x2019;an, Shaanxi, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4695-6134","authenticated-orcid":false,"given":"Nannan","family":"Wang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Integrated Services Networks (ISN), School of Telecommunications Engineering, Xidian University, Xi&#x2019;an, Shaanxi, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7985-0037","authenticated-orcid":false,"given":"Xinbo","family":"Gao","sequence":"additional","affiliation":[{"name":"Chongqing Key Laboratory of Image Cognition, Chongqing University of Posts and Telecommunications, Chongqing, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00453"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00619"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00574"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01565"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.272"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1145\/2897824.2925968"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.167"},{"key":"ref9","first-page":"386","article-title":"Universal style transfer via feature transforms","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"30","author":"Li"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2022.3143356"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref12","article-title":"GANs N\u2019 roses: Stable, controllable, diverse image to image translation (works for videos too!)","author":"Chong","year":"2021","journal-title":"arXiv:2106.06561"},{"key":"ref13","article-title":"U-GAT-IT: Unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation","author":"Kim","year":"2019","journal-title":"arXiv:1907.10830"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2021.3105725"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2022.3145812"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3113786"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref19","first-page":"700","article-title":"Unsupervised image-to-image translation networks","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Liu"},{"key":"ref20","article-title":"Resolution dependent GAN interpolation for controllable image synthesis between domains","author":"Pinkney","year":"2020","journal-title":"arXiv:2010.05334"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1145\/3450626.3459860"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00754"},{"key":"ref23","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. Int. Conf. Mach. Learn.","volume":"139","author":"Radford"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530164"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/3DV.2017.00053"},{"key":"ref26","first-page":"118","article-title":"Visual object networks: Image generation with disentangled 3D representations","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"31","author":"Zhu"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.01008"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00086"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00936"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00446"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2020.3015992"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1145\/3503250"},{"key":"ref33","article-title":"CIPS-3D: A 3D-aware generator of GANs based on conditionally-independent pixel synthesis","author":"Zhou","year":"2021","journal-title":"arXiv:2110.09788"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00580"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01245"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01386"},{"key":"ref37","article-title":"StyleNeRF: A style-based 3D-aware generator for high-resolution image synthesis","author":"Gu","year":"2021","journal-title":"arXiv:2110.08985"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01254"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1145\/3544777"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2019.00038"},{"key":"ref41","first-page":"3481","article-title":"Which training methods for GANs do actually converge?","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Mescheder"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00482"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00232"},{"article-title":"Danbooru2019 portraits: A large-scale anime head illustration dataset","year":"2019","author":"Branwen","key":"ref44"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1145\/3126686.3126736"},{"key":"ref46","article-title":"WebCaricature: A benchmark for caricature recognition","author":"Huo","year":"2017","journal-title":"arXiv:1703.03230"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref48","article-title":"An image is worth one word: Personalizing text-to-image generation using textual inversion","author":"Gal","year":"2022","journal-title":"arXiv:2208.01618"},{"key":"ref49","first-page":"6629","article-title":"GANs trained by a two time-scale update rule converge to a local Nash equilibrium","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Heusel"},{"key":"ref50","article-title":"Demystifying MMD GANs","author":"Bi\u0144kowski","year":"2018","journal-title":"arXiv:1801.01401"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3065230"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2023.3284003"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1145\/3550454.3555437"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1145\/3550454.3555506"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00244"}],"container-title":["IEEE Transactions on Neural Networks and Learning Systems"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/5962385\/11022714\/10979413.pdf?arnumber=10979413","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,4]],"date-time":"2025-06-04T17:57:32Z","timestamp":1749059852000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10979413\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6]]},"references-count":55,"journal-issue":{"issue":"6"},"URL":"https:\/\/doi.org\/10.1109\/tnnls.2025.3559477","relation":{},"ISSN":["2162-237X","2162-2388"],"issn-type":[{"type":"print","value":"2162-237X"},{"type":"electronic","value":"2162-2388"}],"subject":[],"published":{"date-parts":[[2025,6]]}}}