{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:17:24Z","timestamp":1775578644856,"version":"3.50.1"},"reference-count":78,"publisher":"Association for Computing Machinery (ACM)","issue":"4","license":[{"start":{"date-parts":[[2024,7,19]],"date-time":"2024-07-19T00:00:00Z","timestamp":1721347200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"name":"NSFC programs","award":["61976138"],"award-info":[{"award-number":["61976138"]}]},{"name":"NSFC programs","award":["61977047"],"award-info":[{"award-number":["61977047"]}]},{"DOI":"10.13039\/501100003399","name":"STCSM","doi-asserted-by":"crossref","award":["2015F0203-000-06"],"award-info":[{"award-number":["2015F0203-000-06"]}],"id":[{"id":"10.13039\/501100003399","id-type":"DOI","asserted-by":"crossref"}]},{"name":"SHMEC","award":["2019-01-07-00-01-E00003"],"award-info":[{"award-number":["2019-01-07-00-01-E00003"]}]},{"DOI":"10.13039\/501100012166","name":"National Key R&D Program of China","doi-asserted-by":"crossref","award":["2022YFF0902301"],"award-info":[{"award-number":["2022YFF0902301"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Graph."],"published-print":{"date-parts":[[2024,7,19]]},"abstract":"<jats:p>Apparel's significant role in human appearance underscores the importance of garment digitalization for digital human creation. Recent advances in 3D content creation are pivotal for digital human creation. Nonetheless, garment generation from text guidance is still nascent. We introduce a text-driven 3D garment generation framework, DressCode, which aims to democratize design for novices and offer immense potential in fashion design, virtual try-on, and digital human creation. We first introduce SewingGPT, a GPT-based architecture integrating cross-attention with text-conditioned embedding to generate sewing patterns with text guidance. We then tailor a pre-trained Stable Diffusion to generate tile-based Physically-based Rendering (PBR) textures for the garments. By leveraging a large language model, our framework generates CG-friendly garments through natural language interaction. It also facilitates pattern completion and texture editing, streamlining the design process through user-friendly interaction. This framework fosters innovation by allowing creators to freely experiment with designs and incorporate unique elements into their work. With comprehensive evaluations and comparisons with other state-of-the-art methods, our method showcases superior quality and alignment with input prompts. User studies further validate our high-quality rendering results, highlighting its practical utility and potential in production settings. Our project page is https:\/\/IHe-KaiI.github.io\/DressCode\/.<\/jats:p>","DOI":"10.1145\/3658147","type":"journal-article","created":{"date-parts":[[2024,7,19]],"date-time":"2024-07-19T14:47:57Z","timestamp":1721400477000},"page":"1-13","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":33,"title":["DressCode: Autoregressively Sewing and Generating Garments from Text Guidance"],"prefix":"10.1145","volume":"43","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8384-094X","authenticated-orcid":false,"given":"Kai","family":"He","sequence":"first","affiliation":[{"name":"ShanghaiTech University, Shanghai, China"},{"name":"Deemos Technology, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-2056-6057","authenticated-orcid":false,"given":"Kaixin","family":"Yao","sequence":"additional","affiliation":[{"name":"ShanghaiTech University, Shanghai, China"},{"name":"NeuDim, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4837-7152","authenticated-orcid":false,"given":"Qixuan","family":"Zhang","sequence":"additional","affiliation":[{"name":"ShanghaiTech University, Shanghai, China"},{"name":"Deemos Technology, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9198-6853","authenticated-orcid":false,"given":"Jingyi","family":"Yu","sequence":"additional","affiliation":[{"name":"ShanghaiTech University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4301-1474","authenticated-orcid":false,"given":"Lingjie","family":"Liu","sequence":"additional","affiliation":[{"name":"University of Pennsylvania, Philadelphia, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8807-7787","authenticated-orcid":false,"given":"Lan","family":"Xu","sequence":"additional","affiliation":[{"name":"ShanghaiTech University, Shanghai, China"}]}],"member":"320","published-online":{"date-parts":[[2024,7,19]]},"reference":[{"key":"e_1_2_2_1_1","volume-title":"Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al.","author":"Achiam Josh","year":"2023","unstructured":"Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. 2023. Gpt-4 technical report. arXiv preprint arXiv:2303.08774 (2023)."},{"key":"e_1_2_2_2_1","unstructured":"Autodesk INC. 2019. Maya. https:\/\/autodesk.com\/maya"},{"key":"e_1_2_2_3_1","unstructured":"AUTOMATIC1111. 2022. Stable Diffusion Web UI. https:\/\/github.com\/AUTOMATIC1111\/stable-diffusion-webui"},{"key":"e_1_2_2_4_1","volume-title":"Computer Graphics Forum","author":"Bang Seungbae","unstructured":"Seungbae Bang, Maria Korosteleva, and Sung-Hee Lee. 2021. Estimating garment patterns from static scan data. In Computer Graphics Forum, Vol. 40. Wiley Online Library, 273--287."},{"key":"e_1_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/2897824.2925896"},{"key":"e_1_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/2461912.2461975"},{"key":"e_1_2_2_7_1","unstructured":"James Betker Gabriel Goh Li Jing Tim Brooks Jianfeng Wang Linjie Li Long Ouyang Juntang Zhuang Joyce Lee Yufei Guo et al. 2023. Improving image generation with better captions. Computer Science. https:\/\/cdn.openai.com\/papers\/dall-e-3.pdf 2 (2023) 3."},{"key":"e_1_2_2_8_1","unstructured":"Blender Foundation. 2022. Blender. https:\/\/www.blender.org\/"},{"key":"e_1_2_2_9_1","volume-title":"Text2tex: Text-driven texture synthesis via diffusion models. arXiv preprint arXiv:2303.11396","author":"Chen Dave Zhenyu","year":"2023","unstructured":"Dave Zhenyu Chen, Yawar Siddiqui, Hsin-Ying Lee, Sergey Tulyakov, and Matthias Nie\u00dfner. 2023b. Text2tex: Text-driven texture synthesis via diffusion models. arXiv preprint arXiv:2303.11396 (2023)."},{"key":"e_1_2_2_10_1","volume-title":"Fantasia3d: Disentangling geometry and appearance for high-quality text-to-3d content creation. arXiv preprint arXiv:2303.13873","author":"Chen Rui","year":"2023","unstructured":"Rui Chen, Yongwei Chen, Ningxin Jiao, and Kui Jia. 2023a. Fantasia3d: Disentangling geometry and appearance for high-quality text-to-3d content creation. arXiv preprint arXiv:2303.13873 (2023)."},{"key":"e_1_2_2_11_1","first-page":"15147","article-title":"Structure-Preserving 3D Garment Modeling with Neural Sewing Machines","volume":"35","author":"Chen Xipeng","year":"2022","unstructured":"Xipeng Chen, Guangrun Wang, Dizhong Zhu, Xiaodan Liang, Philip Torr, and Liang Lin. 2022. Structure-Preserving 3D Garment Modeling with Neural Sewing Machines. Advances in Neural Information Processing Systems 35 (2022), 15147--15159.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_2_2_12_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/2816795.2818059","article-title":"Garment modeling with a depth camera","volume":"34","author":"Chen Xiaowu","year":"2015","unstructured":"Xiaowu Chen, Bin Zhou, Feixiang Lu, Lin Wang, Lang Bi, and Ping Tan. 2015. Garment modeling with a depth camera. ACM Transactions on Graphics (TOG) 34, 6 (2015), 1--12.","journal-title":"ACM Transactions on Graphics (TOG)"},{"key":"e_1_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/566654.566624"},{"key":"e_1_2_2_14_1","unstructured":"CLO3D. 2024. Marvelous Designer. https:\/\/www.marvelousdesigner.com\/"},{"key":"e_1_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00146"},{"key":"e_1_2_2_16_1","volume-title":"Hyperdiffusion: Generating implicit neural fields with weight-space diffusion. arXiv preprint arXiv:2303.17015","author":"Erko\u00e7 Ziya","year":"2023","unstructured":"Ziya Erko\u00e7, Fangchang Ma, Qi Shan, Matthias Nie\u00dfner, and Angela Dai. 2023. Hyperdiffusion: Generating implicit neural fields with weight-space diffusion. arXiv preprint arXiv:2303.17015 (2023)."},{"key":"e_1_2_2_17_1","volume-title":"Data-driven Garment Pattern Estimation from 3D Geometries. Eurographics 2021-Short Papers","author":"Goto Chihiro","year":"2021","unstructured":"Chihiro Goto and Nobuyuki Umetani. 2021. Data-driven Garment Pattern Estimation from 3D Geometries. Eurographics 2021-Short Papers (2021)."},{"key":"e_1_2_2_18_1","volume-title":"Denoising diffusion probabilistic models. Advances in neural information processing systems 33","author":"Ho Jonathan","year":"2020","unstructured":"Jonathan Ho, Ajay Jain, and Pieter Abbeel. 2020. Denoising diffusion probabilistic models. Advances in neural information processing systems 33 (2020), 6840--6851."},{"key":"e_1_2_2_19_1","volume-title":"Tech: Text-guided reconstruction of lifelike clothed humans. arXiv preprint arXiv:2308.08545","author":"Huang Yangyi","year":"2023","unstructured":"Yangyi Huang, Hongwei Yi, Yuliang Xiu, Tingting Liao, Jiaxiang Tang, Deng Cai, and Justus Thies. 2023. Tech: Text-guided reconstruction of lifelike clothed humans. arXiv preprint arXiv:2308.08545 (2023)."},{"key":"e_1_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00094"},{"key":"e_1_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1002\/cav.1653"},{"key":"e_1_2_2_22_1","volume-title":"Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114","author":"Kingma Diederik P","year":"2013","unstructured":"Diederik P Kingma and Max Welling. 2013. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114 (2013)."},{"key":"e_1_2_2_23_1","volume-title":"Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks, J. Vanschoren and S. Yeung (Eds.)","volume":"1","author":"Korosteleva Maria","year":"2021","unstructured":"Maria Korosteleva and Sung-Hee Lee. 2021. Generating Datasets of 3D Garments with Sewing Patterns. In Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks, J. Vanschoren and S. Yeung (Eds.), Vol. 1. https:\/\/datasets-benchmarks-proceedings.neurips.cc\/paper\/2021\/file\/013d407166ec4fa56eb1e1f8cbe183b9-Paper-round1.pdf"},{"key":"e_1_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530179"},{"key":"e_1_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3618351"},{"key":"e_1_2_2_26_1","volume-title":"International conference on machine learning. PMLR, 12888--12900","author":"Li Junnan","year":"2022","unstructured":"Junnan Li, Dongxu Li, Caiming Xiong, and Steven Hoi. 2022. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In International conference on machine learning. PMLR, 12888--12900."},{"key":"e_1_2_2_27_1","volume-title":"ISP: Multi-Layered Garment Draping with Implicit Sewing Patterns. arXiv preprint arXiv:2305.14100","author":"Li Ren","year":"2023","unstructured":"Ren Li, Beno\u00eet Guillard, and Pascal Fua. 2023b. ISP: Multi-Layered Garment Draping with Implicit Sewing Patterns. arXiv preprint arXiv:2305.14100 (2023)."},{"key":"e_1_2_2_28_1","volume-title":"DiffAvatar: Simulation-Ready Garment Optimization with Differentiable Simulation. arXiv preprint arXiv:2311.12194","author":"Li Yifei","year":"2023","unstructured":"Yifei Li, Hsiao-yu Chen, Egor Larionov, Nikolaos Sarafianos, Wojciech Matusik, and Tuur Stuyck. 2023a. DiffAvatar: Simulation-Ready Garment Optimization with Differentiable Simulation. arXiv preprint arXiv:2311.12194 (2023)."},{"key":"e_1_2_2_29_1","volume-title":"arXiv preprint arXiv:2308.10899","author":"Liao Tingting","year":"2023","unstructured":"Tingting Liao, Hongwei Yi, Yuliang Xiu, Jiaxaing Tang, Yangyi Huang, Justus Thies, and Michael J Black. 2023. Tada! text to animatable digital avatars. arXiv preprint arXiv:2308.10899 (2023)."},{"key":"e_1_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00037"},{"key":"e_1_2_2_31_1","volume-title":"Towards Garment Sewing Pattern Reconstruction from a Single Image. ACM Transactions on Graphics (SIGGRAPH Asia)","author":"Liu Lijuan","year":"2023","unstructured":"Lijuan Liu, Xiangyu Xu, Zhijie Lin, Jiabin Liang, and Shuicheng Yan. 2023d. Towards Garment Sewing Pattern Reconstruction from a Single Image. ACM Transactions on Graphics (SIGGRAPH Asia) (2023)."},{"key":"e_1_2_2_32_1","unstructured":"Minghua Liu Chao Xu Haian Jin Linghao Chen Zexiang Xu Hao Su et al. 2023c. One-2-3-45: Any single image to 3d mesh in 45 seconds without per-shape optimization. arXiv preprint arXiv:2306.16928 (2023)."},{"key":"e_1_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00853"},{"key":"e_1_2_2_34_1","volume-title":"SyncDreamer: Generating Multiview-consistent Images from a Single-view Image. arXiv preprint arXiv:2309.03453","author":"Liu Yuan","year":"2023","unstructured":"Yuan Liu, Cheng Lin, Zijiao Zeng, Xiaoxiao Long, Lingjie Liu, Taku Komura, and Wenping Wang. 2023a. SyncDreamer: Generating Multiview-consistent Images from a Single-view Image. arXiv preprint arXiv:2309.03453 (2023)."},{"key":"e_1_2_2_35_1","doi-asserted-by":"crossref","unstructured":"Xiaoxiao Long Yuan-Chen Guo Cheng Lin Yuan Liu Zhiyang Dou Lingjie Liu Yuexin Ma Song-Hai Zhang Marc Habermann Christian Theobalt et al. 2023. Wonder3d: Single image to 3d using cross-domain diffusion. arXiv preprint arXiv:2310.15008 (2023).","DOI":"10.1109\/CVPR52733.2024.00951"},{"key":"e_1_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01117"},{"key":"e_1_2_2_37_1","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 8446--8455","author":"Melas-Kyriazi Luke","year":"2023","unstructured":"Luke Melas-Kyriazi, Iro Laina, Christian Rupprecht, and Andrea Vedaldi. 2023. Realfusion: 360deg reconstruction of any object from a single image. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 8446--8455."},{"key":"e_1_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01218"},{"key":"e_1_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503250"},{"key":"e_1_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00810"},{"key":"e_1_2_2_41_1","volume-title":"International conference on machine learning. PMLR, 7220--7229","author":"Nash Charlie","year":"2020","unstructured":"Charlie Nash, Yaroslav Ganin, SM Ali Eslami, and Peter Battaglia. 2020. Polygen: An autoregressive generative model of 3d meshes. In International conference on machine learning. PMLR, 7220--7229."},{"key":"e_1_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530145"},{"key":"e_1_2_2_43_1","volume-title":"DreamFusion: Text-to-3D using 2D Diffusion. arXiv","author":"Poole Ben","year":"2022","unstructured":"Ben Poole, Ajay Jain, Jonathan T. Barron, and Ben Mildenhall. 2022. DreamFusion: Text-to-3D using 2D Diffusion. arXiv (2022)."},{"key":"e_1_2_2_44_1","volume-title":"PersonalTailor: Personalizing 2D Pattern Design from 3D Garment Point Clouds. arXiv preprint arXiv:2303.09695","author":"Qi Anran","year":"2023","unstructured":"Anran Qi, Sauradip Nag, Xiatian Zhu, and Ariel Shamir. 2023. PersonalTailor: Personalizing 2D Pattern Design from 3D Garment Point Clouds. arXiv preprint arXiv:2303.09695 (2023)."},{"key":"e_1_2_2_45_1","unstructured":"Guocheng Qian Jinjie Mai Abdullah Hamdi Jian Ren Aliaksandr Siarohin Bing Li Hsin-Ying Lee Ivan Skorokhodov Peter Wonka Sergey Tulyakov et al. 2023. Magic123: One image to high-quality 3d object generation using both 2d and 3d diffusion priors. arXiv preprint arXiv:2306.17843 (2023)."},{"key":"e_1_2_2_46_1","volume-title":"RichDreamer: A Generalizable Normal-Depth Diffusion Model for Detail Richness in Text-to-3D. arXiv preprint arXiv:2311.16918","author":"Qiu Lingteng","year":"2023","unstructured":"Lingteng Qiu, Guanying Chen, Xiaodong Gu, Qi Zuo, Mutian Xu, Yushuang Wu, Weihao Yuan, Zilong Dong, Liefeng Bo, and Xiaoguang Han. 2023. RichDreamer: A Generalizable Normal-Depth Diffusion Model for Detail Richness in Text-to-3D. arXiv preprint arXiv:2311.16918 (2023)."},{"key":"e_1_2_2_47_1","volume-title":"International conference on machine learning. PMLR, 8748--8763","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021. Learning transferable visual models from natural language supervision. In International conference on machine learning. PMLR, 8748--8763."},{"key":"e_1_2_2_48_1","doi-asserted-by":"crossref","unstructured":"Amit Raj Srinivas Kaza Ben Poole Michael Niemeyer Nataniel Ruiz Ben Mildenhall Shiran Zada Kfir Aberman Michael Rubinstein Jonathan Barron et al. 2023. Dreambooth3d: Subject-driven text-to-3d generation. arXiv preprint arXiv:2303.13508 (2023).","DOI":"10.1109\/ICCV51070.2023.00223"},{"key":"e_1_2_2_49_1","volume-title":"Texture: Text-guided texturing of 3d shapes. arXiv preprint arXiv:2302.01721","author":"Richardson Elad","year":"2023","unstructured":"Elad Richardson, Gal Metzer, Yuval Alaluf, Raja Giryes, and Daniel Cohen-Or. 2023. Texture: Text-guided texturing of 3d shapes. arXiv preprint arXiv:2302.01721 (2023)."},{"key":"e_1_2_2_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"e_1_2_2_51_1","volume-title":"U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention-MICCAI 2015: 18th International Conference","author":"Ronneberger Olaf","year":"2015","unstructured":"Olaf Ronneberger, Philipp Fischer, and Thomas Brox. 2015. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention-MICCAI 2015: 18th International Conference, Munich, Germany, October 5--9, 2015, Proceedings, Part III 18. Springer, 234--241."},{"key":"e_1_2_2_52_1","volume-title":"Garment3DGen: 3D Garment Stylization and Texture Generation. arXiv preprint arXiv:2403.18816","author":"Sarafianos Nikolaos","year":"2024","unstructured":"Nikolaos Sarafianos, Tuur Stuyck, Xiaoyu Xiang, Yilei Li, Jovan Popovic, and Rakesh Ranjan. 2024. Garment3DGen: 3D Garment Stylization and Texture Generation. arXiv preprint arXiv:2403.18816 (2024)."},{"key":"e_1_2_2_53_1","volume-title":"Let 2d diffusion model know 3d-consistency for robust text-to-3d generation. arXiv preprint arXiv:2303.07937","author":"Seo Junyoung","year":"2023","unstructured":"Junyoung Seo, Wooseok Jang, Min-Seop Kwak, Jaehoon Ko, Hyeonsu Kim, Junho Kim, Jin-Hwa Kim, Jiyoung Lee, and Seungryong Kim. 2023. Let 2d diffusion model know 3d-consistency for robust text-to-3d generation. arXiv preprint arXiv:2303.07937 (2023)."},{"key":"e_1_2_2_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/3197517.3201356"},{"key":"e_1_2_2_55_1","first-page":"6087","article-title":"Deep marching tetrahedra: a hybrid representation for high-resolution 3d shape synthesis","volume":"34","author":"Shen Tianchang","year":"2021","unstructured":"Tianchang Shen, Jun Gao, Kangxue Yin, Ming-Yu Liu, and Sanja Fidler. 2021. Deep marching tetrahedra: a hybrid representation for high-resolution 3d shape synthesis. Advances in Neural Information Processing Systems 34 (2021), 6087--6101.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_2_2_56_1","volume-title":"Proceedings, Part XVIII 16","author":"Shen Yu","year":"2020","unstructured":"Yu Shen, Junbang Liang, and Ming C Lin. 2020. Gan-based garment generation using sewing pattern images. In Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, August 23--28, 2020, Proceedings, Part XVIII 16. Springer, 225--247."},{"key":"e_1_2_2_57_1","volume-title":"a single image to consistent multi-view diffusion base model. arXiv preprint arXiv:2310.15110","author":"Shi Ruoxi","year":"2023","unstructured":"Ruoxi Shi, Hansheng Chen, Zhuoyang Zhang, Minghua Liu, Chao Xu, Xinyue Wei, Linghao Chen, Chong Zeng, and Hao Su. 2023a. Zero123++: a single image to consistent multi-view diffusion base model. arXiv preprint arXiv:2310.15110 (2023)."},{"key":"e_1_2_2_58_1","volume-title":"Mvdream: Multi-view diffusion for 3d generation. arXiv preprint arXiv:2308.16512","author":"Shi Yichun","year":"2023","unstructured":"Yichun Shi, Peng Wang, Jianglong Ye, Mai Long, Kejie Li, and Xiao Yang. 2023b. Mvdream: Multi-view diffusion for 3d generation. arXiv preprint arXiv:2308.16512 (2023)."},{"key":"e_1_2_2_59_1","volume-title":"MeshGPT: Generating Triangle Meshes with Decoder-Only Transformers. arXiv preprint arXiv:2311.15475","author":"Siddiqui Yawar","year":"2023","unstructured":"Yawar Siddiqui, Antonio Alliegro, Alexey Artemov, Tatiana Tommasi, Daniele Sirigatti, Vladislav Rosov, Angela Dai, and Matthias Nie\u00dfner. 2023. MeshGPT: Generating Triangle Meshes with Decoder-Only Transformers. arXiv preprint arXiv:2311.15475 (2023)."},{"key":"e_1_2_2_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2020.3027763"},{"key":"e_1_2_2_61_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3168569"},{"key":"e_1_2_2_62_1","volume-title":"Make-it-3d: High-fidelity 3d creation from a single image with diffusion prior. arXiv preprint arXiv:2303.14184","author":"Tang Junshu","year":"2023","unstructured":"Junshu Tang, Tengfei Wang, Bo Zhang, Ting Zhang, Ran Yi, Lizhuang Ma, and Dong Chen. 2023. Make-it-3d: High-fidelity 3d creation from a single image with diffusion prior. arXiv preprint arXiv:2303.14184 (2023)."},{"key":"e_1_2_2_63_1","volume-title":"TextMesh: Generation of Realistic 3D Meshes From Text Prompts. arXiv preprint arXiv:2304.12439","author":"Tsalicoglou Christina","year":"2023","unstructured":"Christina Tsalicoglou, Fabian Manhardt, Alessio Tonioni, Michael Niemeyer, and Federico Tombari. 2023. TextMesh: Generation of Realistic 3D Meshes From Text Prompts. arXiv preprint arXiv:2304.12439 (2023)."},{"key":"e_1_2_2_64_1","doi-asserted-by":"publisher","DOI":"10.1145\/2010324.1964985"},{"key":"e_1_2_2_65_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01214"},{"key":"e_1_2_2_66_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3272127.3275074","article-title":"Learning a shared shape space for multimodal garment design","volume":"37","author":"Wang Tuanfeng Y","year":"2018","unstructured":"Tuanfeng Y Wang, Duygu Ceylan, Jovan Popovi\u0107, and Niloy J Mitra. 2018. Learning a shared shape space for multimodal garment design. ACM Transactions on Graphics 37, 6 (2018), 1--13.","journal-title":"ACM Transactions on Graphics"},{"key":"e_1_2_2_67_1","volume-title":"ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation. arXiv preprint arXiv:2305.16213","author":"Wang Zhengyi","year":"2023","unstructured":"Zhengyi Wang, Cheng Lu, Yikai Wang, Fan Bao, Chongxuan Li, Hang Su, and Jun Zhu. 2023b. ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation. arXiv preprint arXiv:2305.16213 (2023)."},{"key":"e_1_2_2_68_1","doi-asserted-by":"publisher","DOI":"10.1145\/3610548.3618168"},{"key":"e_1_2_2_69_1","unstructured":"Dejia Xu Yifan Jiang Peihao Wang Zhiwen Fan Yi Wang and Zhangyang Wang."},{"key":"e_1_2_2_70_1","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 4479--4489","unstructured":"2023. NeuralLift-360: Lifting an In-the-Wild 2D Photo to a 3D Object With 360deg Views. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 4479--4489."},{"key":"e_1_2_2_71_1","doi-asserted-by":"publisher","DOI":"10.1145\/3026479"},{"key":"e_1_2_2_72_1","volume-title":"Consistent-1-to-3: Consistent image to 3d view synthesis via geometry-aware diffusion models. arXiv preprint arXiv:2310.03020","author":"Ye Jianglong","year":"2023","unstructured":"Jianglong Ye, Peng Wang, Kejie Li, Yichun Shi, and Heng Wang. 2023. Consistent-1-to-3: Consistent image to 3d view synthesis via geometry-aware diffusion models. arXiv preprint arXiv:2310.03020 (2023)."},{"key":"e_1_2_2_73_1","unstructured":"Zhengming Yu Zhiyang Dou Xiaoxiao Long Cheng Lin Zekun Li Yuan Liu Norman M\u00fcller Taku Komura Marc Habermann Christian Theobalt et al. 2023. Surf-D: High-Quality Surface Generation for Arbitrary Topologies using Diffusion Models. arXiv preprint arXiv:2311.17050 (2023)."},{"key":"e_1_2_2_74_1","volume-title":"DreamFace: Progressive Generation of Animatable 3D Faces under Text Guidance. arXiv preprint arXiv:2304.03117","author":"Zhang Longwen","year":"2023","unstructured":"Longwen Zhang, Qiwei Qiu, Hongyang Lin, Qixuan Zhang, Cheng Shi, Wei Yang, Ye Shi, Sibei Yang, Lan Xu, and Jingyi Yu. 2023a. DreamFace: Progressive Generation of Animatable 3D Faces under Text Guidance. arXiv preprint arXiv:2304.03117 (2023)."},{"key":"e_1_2_2_75_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"e_1_2_2_76_1","volume-title":"EfficientDreamer: High-Fidelity and Robust 3D Creation via Orthogonal-view Diffusion Prior. arXiv preprint arXiv:2308.13223","author":"Zhao Minda","year":"2023","unstructured":"Minda Zhao, Chaoyi Zhao, Xinyue Liang, Lincheng Li, Zeng Zhao, Zhipeng Hu, Changjie Fan, and Xin Yu. 2023. EfficientDreamer: High-Fidelity and Robust 3D Creation via Orthogonal-view Diffusion Prior. arXiv preprint arXiv:2308.13223 (2023)."},{"key":"e_1_2_2_77_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3618309","article-title":"GroomGen: A High-Quality Generative Hair Model Using Hierarchical Latent Representations","volume":"42","author":"Zhou Yuxiao","year":"2023","unstructured":"Yuxiao Zhou, Menglei Chai, Alessandro Pepe, Markus Gross, and Thabo Beeler. 2023. GroomGen: A High-Quality Generative Hair Model Using Hierarchical Latent Representations. ACM Transactions on Graphics (TOG) 42, 6 (2023), 1--16.","journal-title":"ACM Transactions on Graphics (TOG)"},{"key":"e_1_2_2_78_1","volume-title":"Proceedings, Part I 16","author":"Zhu Heming","year":"2020","unstructured":"Heming Zhu, Yu Cao, Hang Jin, Weikai Chen, Dong Du, Zhangye Wang, Shuguang Cui, and Xiaoguang Han. 2020. Deep fashion3d: A dataset and benchmark for 3d garment reconstruction from single images. In Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, August 23--28, 2020, Proceedings, Part I 16. Springer, 512--530."}],"container-title":["ACM Transactions on Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3658147","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3658147","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:05:54Z","timestamp":1750291554000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3658147"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,19]]},"references-count":78,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,7,19]]}},"alternative-id":["10.1145\/3658147"],"URL":"https:\/\/doi.org\/10.1145\/3658147","relation":{},"ISSN":["0730-0301","1557-7368"],"issn-type":[{"value":"0730-0301","type":"print"},{"value":"1557-7368","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7,19]]},"assertion":[{"value":"2024-07-19","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}