{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T11:58:19Z","timestamp":1770724699863,"version":"3.49.0"},"reference-count":47,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100004663","name":"Ministry of Science and Technology, Taiwan","doi-asserted-by":"publisher","award":["108-2634-F-002-018"],"award-info":[{"award-number":["108-2634-F-002-018"]}],"id":[{"id":"10.13039\/501100004663","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100005144","name":"Qualcomm","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100005144","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. on Image Process."],"published-print":{"date-parts":[[2020]]},"DOI":"10.1109\/tip.2019.2952707","type":"journal-article","created":{"date-parts":[[2019,11,15]],"date-time":"2019-11-15T21:09:38Z","timestamp":1573852178000},"page":"2795-2807","source":"Crossref","is-referenced-by-count":13,"title":["A Multi-Domain and Multi-Modal Representation Disentangler for Cross-Domain Image Manipulation and Classification"],"prefix":"10.1109","volume":"29","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0102-7101","authenticated-orcid":false,"given":"Fu-En","family":"Yang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8495-9053","authenticated-orcid":false,"given":"Jing-Cheng","family":"Chang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1792-9978","authenticated-orcid":false,"given":"Chung-Chi","family":"Tsai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2333-157X","authenticated-orcid":false,"given":"Yu-Chiang Frank","family":"Wang","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"2014","journal-title":"arXiv 1412 6980"},{"key":"ref38","first-page":"1","article-title":"Automatic differentiation in pytorch","author":"paszke","year":"2017","journal-title":"Proc NIPS"},{"key":"ref33","first-page":"1857","article-title":"Learning to discover cross-domain relations with generative adversarial networks","author":"kim","year":"2017","journal-title":"Proc 34th Int Conf Mach Learn"},{"key":"ref32","article-title":"Auto-encoding variational bayes","author":"kingma","year":"2013","journal-title":"arXiv 1312 6114"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2018\/131"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.180"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2018.2874567"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01216-8_3"},{"key":"ref35","article-title":"AttGAN: Facial attribute editing by only changing what you want","author":"he","year":"2017","journal-title":"arXiv 1711 10678"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.310"},{"key":"ref10","first-page":"700","article-title":"Unsupervised image-to-image translation networks","author":"liu","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.425"},{"key":"ref11","first-page":"1180","article-title":"Unsupervised domain adaptation by backpropagation","author":"ganin","year":"2015","journal-title":"Proc 32nd Int Conf Mach Learn"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.316"},{"key":"ref13","first-page":"343","article-title":"Domain separation networks","author":"bousmalis","year":"2016","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.301"},{"key":"ref15","article-title":"CyCADA: Cycle-consistent adversarial domain adaptation","author":"hoffman","year":"2017","journal-title":"arXiv 1711 03213"},{"key":"ref16","first-page":"2172","article-title":"InfoGAN: Interpretable representation learning by information maximizing generative adversarial nets","author":"chen","year":"2016","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref17","first-page":"1","article-title":"$\\beta$\n-VAE: Learning basic visual concepts with a constrained variational framework","author":"higgins","year":"2017","journal-title":"Proc Int Conf Learn Represent (ICLR)"},{"key":"ref18","first-page":"2642","article-title":"Conditional image synthesis with auxiliary classifier GANs","author":"odena","year":"2017","journal-title":"Proc 34th Int Conf Mach Learn"},{"key":"ref19","article-title":"Detach and adapt: Learning cross-domain disentangled deep representation","author":"liu","year":"2017","journal-title":"arXiv 1705 01314"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.141"},{"key":"ref4","first-page":"386","article-title":"Universal style transfer via feature transforms","author":"li","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref27","article-title":"MoCoGAN: Decomposing motion and content for video generation","author":"tulyakov","year":"2017","journal-title":"arXiv 1707 04993"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.167"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00394"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01237-3_43"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref7","first-page":"1","article-title":"Unsupervised cross-domain image generation","author":"taigman","year":"2017","journal-title":"Proc Int Conf Learn Represent (ICLR)"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"ref9","first-page":"469","article-title":"Coupled generative adversarial networks","author":"liu","year":"2016","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref1","article-title":"A neural algorithm of artistic style","author":"gatys","year":"2015","journal-title":"arXiv 1508 06576"},{"key":"ref46","article-title":"Generate To adapt: Aligning domains using generative adversarial networks","author":"sankaranarayanan","year":"2017","journal-title":"arXiv 1704 01705"},{"key":"ref20","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46493-0_36"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01246-5_3"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00845"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01219-9_11"},{"key":"ref42","first-page":"6626","article-title":"GANs trained by a two time-scale update rule converge to a local nash equilibrium","author":"heusel","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref24","first-page":"465","article-title":"Toward multimodal image-to-image translation","author":"zhu","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00122"},{"key":"ref23","article-title":"StarGAN: Unified generative adversarial networks for multi-domain image-to-image translation","author":"choi","year":"2017","journal-title":"arXiv 1711 09020"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.18"},{"key":"ref26","article-title":"A unified feature disentangler for multi-domain image translation and manipulation","author":"liu","year":"2018","journal-title":"arXiv 1809 01361"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"ref25","first-page":"1287","article-title":"Image-to-image translation for cross-domain disentanglement","author":"gonzalez-garcia","year":"2018","journal-title":"Proc Adv Neural Inf Process Syst"}],"container-title":["IEEE Transactions on Image Processing"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/83\/8835130\/08902223.pdf?arnumber=8902223","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,4,27]],"date-time":"2022-04-27T14:38:51Z","timestamp":1651070331000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8902223\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"references-count":47,"URL":"https:\/\/doi.org\/10.1109\/tip.2019.2952707","relation":{},"ISSN":["1057-7149","1941-0042"],"issn-type":[{"value":"1057-7149","type":"print"},{"value":"1941-0042","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]}}}