{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T19:51:38Z","timestamp":1772913098751,"version":"3.50.1"},"reference-count":41,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"8","license":[{"start":{"date-parts":[[2025,8,1]],"date-time":"2025-08-01T00:00:00Z","timestamp":1754006400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2025,8,1]],"date-time":"2025-08-01T00:00:00Z","timestamp":1754006400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,8,1]],"date-time":"2025-08-01T00:00:00Z","timestamp":1754006400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Guangdong International Technology Cooperation Project","award":["2022A0505050009"],"award-info":[{"award-number":["2022A0505050009"]}]},{"name":"China National Key R&#x0026;D Program","award":["2023YFE0202700"],"award-info":[{"award-number":["2023YFE0202700"]}]},{"name":"Key-Area Research and Development Program of Guangzhou City","award":["2023B01J0022"],"award-info":[{"award-number":["2023B01J0022"]}]},{"name":"Guangdong Natural Science Funds for Distinguished Young Scholar","award":["2023B1515020097"],"award-info":[{"award-number":["2023B1515020097"]}]},{"name":"Singapore MOE Tier 1 Funds","award":["MSS23C002"],"award-info":[{"award-number":["MSS23C002"]}]},{"name":"National Research Foundation Singapore under the AI Singapore Programme","award":["AISG3-GV-2023-011"],"award-info":[{"award-number":["AISG3-GV-2023-011"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Visual. Comput. Graphics"],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1109\/tvcg.2024.3397712","type":"journal-article","created":{"date-parts":[[2024,5,7]],"date-time":"2024-05-07T17:28:14Z","timestamp":1715102894000},"page":"4198-4209","source":"Crossref","is-referenced-by-count":6,"title":["DreamAnime: Learning Style-Identity Textual Disentanglement for Anime and Beyond"],"prefix":"10.1109","volume":"31","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0377-8646","authenticated-orcid":false,"given":"Chenshu","family":"Xu","sequence":"first","affiliation":[{"name":"School of Computing and Information Systems, Singapore Management University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3383-4349","authenticated-orcid":false,"given":"Yangyang","family":"Xu","sequence":"additional","affiliation":[{"name":"Department of Computer Science, University of Hong Kong, Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7662-9831","authenticated-orcid":false,"given":"Huaidong","family":"Zhang","sequence":"additional","affiliation":[{"name":"School of Future Technology, South China University of Technology, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8006-3663","authenticated-orcid":false,"given":"Xuemiao","family":"Xu","sequence":"additional","affiliation":[{"name":"School of Computer Science and Engineering, South China University of Technology, Guangzhou, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3802-4644","authenticated-orcid":false,"given":"Shengfeng","family":"He","sequence":"additional","affiliation":[{"name":"School of Computing and Information Systems, Singapore Management University, Singapore"}]}],"member":"263","reference":[{"key":"ref1","first-page":"6840","article-title":"Denoising diffusion probabilistic models","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Ho"},{"key":"ref2","article-title":"Diffusion models already have a semantic latent space","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Kwon"},{"key":"ref3","first-page":"8780","article-title":"Diffusion models beat GANs on image synthesis","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Dhariwal"},{"key":"ref4","article-title":"Hierarchical text-conditional image generation with clip latents","author":"Ramesh","year":"2022"},{"key":"ref5","article-title":"An image is worth one word: Personalizing text-to-image generation using textual inversion","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Gal"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-72633-0_7"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.5555\/2969033.2969125"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00453"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"ref11","article-title":"Progressive growing of GANs for improved quality, stability, and variation","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Karras"},{"key":"ref12","article-title":"Auto-encoding variational bayes","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Kingma"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"ref14","first-page":"14866","article-title":"Generating diverse high-fidelity images with VQ-VAE-2","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Razavi"},{"key":"ref15","article-title":"Nice: Non-linear independent components estimation","volume-title":"Proc. Int. Conf. Learn. Representations WorkShop","author":"Dinh"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00192"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01461"},{"key":"ref19","first-page":"1018","article-title":"Explicit disentanglement of appearance and perspective in generative models","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Skafte"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00189"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1145\/3610548.3618154"},{"key":"ref22","article-title":"DAF: Re: A challenging, crowd-sourced, large-scale, long-tailed dataset for anime character recognition","author":"Rios","year":"2021"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ISCAS48785.2022.9937519"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1145\/3585002"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3113786"},{"key":"ref26","article-title":"U-GAT-IT: Unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Kim"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1145\/3240508.3240661"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/3306214.3338560"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2024.3357568"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2023.3264883"},{"key":"ref31","article-title":"Diffusers: State-of-the-art diffusion models","author":"von Platen","year":"2022"},{"key":"ref32","first-page":"8026","article-title":"PyTorch: An imperative style, high-performance deep learning library","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Paszke"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1145\/3588432.3591506"},{"key":"ref34","article-title":"BLIP-diffusion: Pre-trained subject representation for controllable text-to-image generation and editing","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Li"},{"key":"ref35","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Radford"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00209"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00978"},{"key":"ref39","article-title":"SDEDIT: Guided image synthesis and editing with stochastic differential equations","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Meng"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19787-1_8"},{"key":"ref41","first-page":"29710","article-title":"BlendGAN: Implicitly GAN blending for arbitrary stylized face generation","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Liu"}],"container-title":["IEEE Transactions on Visualization and Computer Graphics"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/2945\/11063255\/10521816.pdf?arnumber=10521816","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T05:13:36Z","timestamp":1751519616000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10521816\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8]]},"references-count":41,"journal-issue":{"issue":"8"},"URL":"https:\/\/doi.org\/10.1109\/tvcg.2024.3397712","relation":{},"ISSN":["1077-2626","1941-0506","2160-9306"],"issn-type":[{"value":"1077-2626","type":"print"},{"value":"1941-0506","type":"electronic"},{"value":"2160-9306","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,8]]}}}