{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T18:06:43Z","timestamp":1775326003277,"version":"3.50.1"},"reference-count":50,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"12","license":[{"start":{"date-parts":[[2023,12,1]],"date-time":"2023-12-01T00:00:00Z","timestamp":1701388800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2023,12,1]],"date-time":"2023-12-01T00:00:00Z","timestamp":1701388800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,12,1]],"date-time":"2023-12-01T00:00:00Z","timestamp":1701388800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Research Grants Council of the Hong Kong Special Administrative Region","award":["CUHK 14206320"],"award-info":[{"award-number":["CUHK 14206320"]}]},{"name":"General Research Fund of Hong Kong","award":["17202422"],"award-info":[{"award-number":["17202422"]}]},{"name":"Hong Kong Research Grant Council - Early Career Scheme","award":["27209621"],"award-info":[{"award-number":["27209621"]}]},{"name":"General Research Fund","award":["17202422"],"award-info":[{"award-number":["17202422"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62202151"],"award-info":[{"award-number":["62202151"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2023,12]]},"DOI":"10.1109\/tpami.2023.3321329","type":"journal-article","created":{"date-parts":[[2023,10,2]],"date-time":"2023-10-02T17:52:54Z","timestamp":1696269174000},"page":"14385-14403","source":"Crossref","is-referenced-by-count":10,"title":["DreamStone: Image as a Stepping Stone for Text-Guided 3D Shape Generation"],"prefix":"10.1109","volume":"45","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8434-3224","authenticated-orcid":false,"given":"Zhengzhe","family":"Liu","sequence":"first","affiliation":[{"name":"Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9538-5879","authenticated-orcid":false,"given":"Peng","family":"Dai","sequence":"additional","affiliation":[{"name":"Department of Electrical and Electronic Engineering, University of Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4266-6420","authenticated-orcid":false,"given":"Ruihui","family":"Li","sequence":"additional","affiliation":[{"name":"College of Computer Science and Electronic Engineering, Hunan University, Changsha, Hunan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4285-1626","authenticated-orcid":false,"given":"Xiaojuan","family":"Qi","sequence":"additional","affiliation":[{"name":"Department of Electrical and Electronic Engineering, University of Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5238-593X","authenticated-orcid":false,"given":"Chi-Wing","family":"Fu","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong"}]}],"member":"263","reference":[{"key":"ref13","article-title":"Controllable text-to-image generation","author":"li","year":"2019","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref12","article-title":"ClipMatrix: Text-controlled creation of 3D textured meshes","author":"jetchev","year":"2021"},{"key":"ref15","article-title":"Mind the gap: Understanding the modality gap in multi-modal contrastive representation learning","author":"liang","year":"2022","journal-title":"Proc NeurIPS"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00790"},{"key":"ref11","first-page":"857","article-title":"Zero-shot text-guided object generation with drefam fields","author":"jain","year":"2022","journal-title":"Proc IEEE Conf Comput Vis Pattern Recognit"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.142619"},{"key":"ref17","article-title":"ISS: Image as stetting stone for text-guided 3D shape generation","author":"liu","year":"2023","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref16","article-title":"FuseDream: Training-free text-to-image generation with improved CLIP+GAN space optimization","author":"liu","year":"2021"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01313"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01737"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-79150-6_25"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2019.2953753"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00143"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2018.2856256"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.629"},{"key":"ref42","article-title":"CLIP-GEN: Language-free training of a text-to-image generator with CLIP","author":"wang","year":"2022"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475226"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00229"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/ICME46284.2020.9102904"},{"key":"ref49","first-page":"17886","article-title":"LAFITE: Towards language-free training for text-to-image generation","author":"zhou","year":"2022","journal-title":"Proc IEEE Conf Comput Vis Pattern Recognit"},{"key":"ref8","article-title":"GANs trained by a two time-scale update rule converge to a local Nash equilibrium","author":"heusel","year":"2017","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref7","article-title":"Get3D: A generative model of high quality 3D textured shapes learned from images","author":"gao","year":"2022","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530094"},{"key":"ref4","first-page":"100","article-title":"Text2Shape: Generating shapes from natural language by learning joint embeddings","author":"chen","year":"2018","journal-title":"Proc Asian Conf Comput Vis"},{"key":"ref3","article-title":"ShapeNet: An information-rich 3D model repository","author":"chang","year":"2015"},{"key":"ref6","article-title":"CogView: Mastering text-to-image generation via transformers","author":"ding","year":"2021","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref5","article-title":"TANGO: Text-driven photorealistic and robust 3D stylization via lighting decomposition","author":"chen","year":"2022","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00381"},{"key":"ref35","article-title":"Photorealistic text-to-image diffusion models with deep language understanding","author":"saharia","year":"2022","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref34","article-title":"Network-to-network translation with conditional invertible neural networks","author":"rombach","year":"2020","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN48605.2020.9207584"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01805"},{"key":"ref31","article-title":"Learning what and where to draw","author":"reed","year":"2016","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref30","article-title":"Generative adversarial text to image synthesis","author":"reed","year":"2016","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01072"},{"key":"ref2","first-page":"3763","article-title":"Pre-train, self-train, distill: A simple recipe for supersizing 3D reconstruction","author":"alwala","year":"2022","journal-title":"Proc IEEE Conf Comput Vis Pattern Recognit"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/3DV50981.2020.00080"},{"key":"ref39","article-title":"Stable-DreamFusion: Text-to-3D with stable-diffusion","author":"tang","year":"2022"},{"key":"ref38","article-title":"Conditional image generation and manipulation for user-specified content","author":"stap","year":"2020","journal-title":"Proc IEEE Conf Comp Vis Pattern Recognit"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00209"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00356"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00160"},{"key":"ref25","article-title":"DreamFusion: Text-to-3D using 2D diffusion","author":"poole","year":"2023","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1145\/3550469.3555392"},{"key":"ref22","article-title":"GLIDE: Towards photorealistic image generation and editing with text-guided diffusion models","author":"nichol","year":"2022","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00810"},{"key":"ref28","article-title":"Hierarchical text-conditional image generation with CLIP latents","author":"ramesh","year":"2022"},{"key":"ref27","article-title":"Learning transferable visual models from natural language supervision","author":"radford","year":"2021","journal-title":"Proc Int Conf Mach Learn"},{"key":"ref29","article-title":"Zero-shot text-to-image generation","author":"ramesh","year":"2021","journal-title":"Proc Int Conf Mach Learn"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/34\/10308548\/10269027.pdf?arnumber=10269027","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,27]],"date-time":"2023-11-27T19:48:28Z","timestamp":1701114508000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10269027\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12]]},"references-count":50,"journal-issue":{"issue":"12"},"URL":"https:\/\/doi.org\/10.1109\/tpami.2023.3321329","relation":{},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"value":"0162-8828","type":"print"},{"value":"2160-9292","type":"electronic"},{"value":"1939-3539","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12]]}}}