{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T16:46:28Z","timestamp":1776098788377,"version":"3.50.1"},"reference-count":56,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"1","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100012325","name":"NSSFC","doi-asserted-by":"publisher","award":["22ZD05"],"award-info":[{"award-number":["22ZD05"]}],"id":[{"id":"10.13039\/501100012325","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012325","name":"NSSFC","doi-asserted-by":"publisher","award":["62472178"],"award-info":[{"award-number":["62472178"]}],"id":[{"id":"10.13039\/501100012325","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100007219","name":"Natural Science Foundation of Shanghai Municipality","doi-asserted-by":"publisher","award":["24ZR1418300"],"award-info":[{"award-number":["24ZR1418300"]}],"id":[{"id":"10.13039\/100007219","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Visual. Comput. Graphics"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1109\/tvcg.2025.3633875","type":"journal-article","created":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T18:48:04Z","timestamp":1764701284000},"page":"331-341","source":"Crossref","is-referenced-by-count":2,"title":["Sel3DCraft: Interactive Visual Prompts for User-Friendly Text-to-3D Generation"],"prefix":"10.1109","volume":"32","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-8478-2391","authenticated-orcid":false,"given":"Nan","family":"Xiang","sequence":"first","affiliation":[{"name":"East China Normal University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8372-8379","authenticated-orcid":false,"given":"Tianyi","family":"Liang","sequence":"additional","affiliation":[{"name":"East China Normal University, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-9443-3855","authenticated-orcid":false,"given":"Haiwen","family":"Huang","sequence":"additional","affiliation":[{"name":"East China Normal University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8175-874X","authenticated-orcid":false,"given":"Shiqi","family":"Jiang","sequence":"additional","affiliation":[{"name":"East China Normal University, China"}]},{"given":"Hao","family":"Huang","sequence":"additional","affiliation":[{"name":"East China Normal University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3077-0175","authenticated-orcid":false,"given":"Yifei","family":"Huang","sequence":"additional","affiliation":[{"name":"East China Normal University, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-0243-3613","authenticated-orcid":false,"given":"Liangyu","family":"Chen","sequence":"additional","affiliation":[{"name":"East China Normal University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8940-6418","authenticated-orcid":false,"given":"Changbo","family":"Wang","sequence":"additional","affiliation":[{"name":"East China Normal University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9835-2650","authenticated-orcid":false,"given":"Chenhui","family":"Li","sequence":"additional","affiliation":[{"name":"East China Normal University, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1145\/3610548.3618228"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.12310"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.12310"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/tvcg.2022.3209425"},{"issue":"3","key":"ref5","first-page":"8","article-title":"Improving image generation with better captions","volume":"2","author":"Betker","year":"2023","journal-title":"Computer Science"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1145\/3586183.3606725"},{"key":"ref7","article-title":"Language models are few-shot learners","volume-title":"Advances in Neural Information Processing Systems 33: Annual Conference on Neural Infor-mation Processing Systems 2020, NeurIPS 2020","author":"Brown","year":"2020"},{"key":"ref8","article-title":"Mllm-as-a-judge: Assessing multimodal llm-as-a-judge with vision-language benchmark","volume-title":"Forty-first International Conference on Machine Learning, ICML 2024","author":"Chen","year":"2024"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-72691-0_8"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1145\/2617588"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1145\/3586183.3606777"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3501819"},{"key":"ref13","article-title":"Objaverse-xl: A universe of 10m+ 3d objects","volume-title":"Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023","author":"Deitke","year":"2023"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/tvcg.2023.3327168"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.295"},{"key":"ref16","article-title":"REPARO: compositional 3d assets generation with differentiable 3d layout alignment","author":"Han","year":"2024","journal-title":"CoRR, abs\/2405.18525"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-73242-3_26"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.595"},{"key":"ref19","article-title":"LRM: large reconstruction model for single image to 3d","volume-title":"The Twelfth International Conference on Learning Representations, ICLR 2024","author":"Hong","year":"2024"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr.2007.383267"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/tvcg.2023.3245609"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/iccv51070.2023.00371"},{"key":"ref23","first-page":"12888","article-title":"BLIP: bootstrapping language-image pre-training for unified vision-language understanding and generation","volume-title":"International Conference on Machine Learning, ICML 2022,17\u201323 July 2022","volume":"162","author":"Li","year":"2022"},{"key":"ref24","article-title":"Instant3d: Fast text-to-3d with sparse-view generation and large reconstruction model","volume-title":"The Twelfth International Conference on Learning Representations, ICLR 2024","author":"Li","year":"2024"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/tvcg.2024.3388514"},{"key":"ref26","article-title":"Openshape: Scaling up 3d shape representation towards open-world understanding","volume-title":"Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023","author":"Liu","year":"2023"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/iccv51070.2023.00853"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3501825"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1145\/3563657.3596098"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/258734.258887"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1016\/j.patter.2020.100141"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_24"},{"key":"ref33","article-title":"Prompt engineering for text-based generative art","author":"Oppenlaender","year":"2022","journal-title":"CoRR, abs\/2204.13988"},{"key":"ref34","author":"Ouyang","year":"2022","journal-title":"Training language models to follow instructions with human feedback"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1145\/3592436"},{"key":"ref36","author":"Podell","year":"2024","journal-title":"SDXL: improving latent diffusion models for high-resolution image synthesis"},{"key":"ref37","article-title":"Dreamfusion: Text-to-3d using 2d diffusion","volume-title":"The Eleventh International Conference on Learning Representations, ICLR 2023","author":"Poole","year":"2023"},{"key":"ref38","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proceedings of the 38th International Conference on Machine Learning, ICML 2021","volume":"139","author":"Radford","year":"2021"},{"key":"ref39","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proceedings of the 38th International Conference on Machine Learning, ICML 2021","volume":"139","author":"Radford","year":"2021"},{"key":"ref40","article-title":"Hier-archical text-conditional image generation with CLIP latents","author":"Ramesh","year":"2022","journal-title":"CoRR, abs\/2204.06125"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr52688.2022.01042"},{"key":"ref42","article-title":"Mvdream: Multi-view diffusion for 3d generation","volume-title":"The Twelfth International Conference on Learning Representations, ICLR 2024","author":"Shi","year":"2024"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.346"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1145\/3613904.3642847"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/tvcg.2022.3209479"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376593"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-73235-5_1"},{"key":"ref48","article-title":"Dreamgaussian: Generative gaussian splatting for efficient 3d content creation","volume-title":"The Twelfth International Conference on Learning Representations, ICLR 2024","author":"Tang","year":"2024"},{"key":"ref49","article-title":"Volumed-iffusion: Flexible text-to-3d generation with efficient volumetric encoder","author":"Tang","year":"2023","journal-title":"CoRR, abs\/2312.11459"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/tcom.1972.1091157"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1145\/3586182.3616707"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr52729.2023.01214"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.51"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1145\/3491101.3519729"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr52733.2024.02098"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/tvcg.2021.3128157"}],"container-title":["IEEE Transactions on Visualization and Computer Graphics"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/2945\/11373125\/11271758.pdf?arnumber=11271758","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T00:33:24Z","timestamp":1770683604000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11271758\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1]]},"references-count":56,"journal-issue":{"issue":"1"},"URL":"https:\/\/doi.org\/10.1109\/tvcg.2025.3633875","relation":{},"ISSN":["1077-2626","1941-0506","2160-9306"],"issn-type":[{"value":"1077-2626","type":"print"},{"value":"1941-0506","type":"electronic"},{"value":"2160-9306","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1]]}}}