{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T18:55:21Z","timestamp":1774637721236,"version":"3.50.1"},"reference-count":60,"publisher":"Association for Computing Machinery (ACM)","issue":"4","funder":[{"name":"Key R&D Program of Zhejiang","award":["2024C01069"],"award-info":[{"award-number":["2024C01069"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Graph."],"published-print":{"date-parts":[[2025,8,1]]},"abstract":"<jats:p>\n                    Current 3D asset creation pipelines typically consist of three stages: creating multi-view concept art, producing 3D meshes based on the artwork, and painting textures for the meshes\u2014an often labor-intensive process. Automated texture generation offers significant acceleration, but prior methods, which fine-tune 2D diffusion models with multi-view input images, often fail to preserve pixel-level details. These methods primarily emphasize semantic and subject consistency, which do not meet the requirements of artwork-guided texture workflows. To address this, we present\n                    <jats:bold>AlignTex<\/jats:bold>\n                    , a novel framework for generating high-quality textures from 3D meshes and multi-view artwork, ensuring both appearance detail and geometric consistency. AlignTex operates in two stages: aligned image generation and texture refinement. The core of our approach,\n                    <jats:bold>AlignNet<\/jats:bold>\n                    , resolves complex misalignments by extracting information from both the artwork and the mesh, generating images compatible with orthographic projection while maintaining geometric and visual fidelity. After projecting aligned images into the texture space, further refinement addresses seams and self-occlusion using an inpainting model and a geometry-aware texture dilation method. Experimental results demonstrate that AlignTex outperforms baseline methods in generation quality and efficiency, offering a practical solution to enhance 3D asset creation in gaming and film production.\n                  <\/jats:p>","DOI":"10.1145\/3731158","type":"journal-article","created":{"date-parts":[[2025,7,27]],"date-time":"2025-07-27T04:02:22Z","timestamp":1753588942000},"page":"1-12","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["AlignTex: Pixel-Precise Texture Generation from Multi-view Artwork"],"prefix":"10.1145","volume":"44","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8512-0551","authenticated-orcid":false,"given":"Yuqing","family":"Zhang","sequence":"first","affiliation":[{"name":"Zhejiang University, Hangzhou, China"},{"name":"State Key Laboratory of CAD&amp;CG, Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5690-367X","authenticated-orcid":false,"given":"Hao","family":"Xu","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"},{"name":"State Key Laboratory of CAD&amp;CG, Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2432-809X","authenticated-orcid":false,"given":"Yiqian","family":"Wu","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"},{"name":"State Key Laboratory of CAD&amp;CG, Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-6518-9961","authenticated-orcid":false,"given":"Sirui","family":"Chen","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-4968-7306","authenticated-orcid":false,"given":"Sirui","family":"Lin","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-1049-4399","authenticated-orcid":false,"given":"Xiang","family":"Li","sequence":"additional","affiliation":[{"name":"Shenzhen University, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9348-526X","authenticated-orcid":false,"given":"Xifeng","family":"Gao","sequence":"additional","affiliation":[{"name":"Lightspeed Studios, Tencent America, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7339-2920","authenticated-orcid":false,"given":"Xiaogang","family":"Jin","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"},{"name":"State Key Laboratory of CAD&amp;CG, Zhejiang University, Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2025,7,27]]},"reference":[{"key":"e_1_2_2_1_1","unstructured":"Adobe. 2024. substance3d. https:\/\/www.adobe.com\/products\/substance3d.html"},{"key":"e_1_2_2_2_1","unstructured":"Autodesk. 2024a. 3ds Max. https:\/\/www.autodesk.com\/products\/3ds-max"},{"key":"e_1_2_2_3_1","unstructured":"Autodesk. 2024b. Maya. https:\/\/www.autodesk.com\/products\/maya"},{"key":"e_1_2_2_4_1","volume-title":"Meta 3d texturegen: Fast and consistent texture generation for 3d objects. arXiv preprint arXiv:2407.02430","author":"Bensadoun Raphael","year":"2024","unstructured":"Raphael Bensadoun, Yanir Kleiman, Idan Azuri, Omri Harosh, Andrea Vedaldi, Natalia Neverova, and Oran Gafni. 2024. Meta 3d texturegen: Fast and consistent texture generation for 3d objects. arXiv preprint arXiv:2407.02430 (2024)."},{"key":"e_1_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00385"},{"key":"e_1_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01701"},{"key":"e_1_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.02033"},{"key":"e_1_2_2_8_1","volume-title":"Annual Conference on Neural Information Processing Systems 2023","author":"Chen Wenhu","year":"2023","unstructured":"Wenhu Chen, Hexiang Hu, Yandong Li, Nataniel Ruiz, Xuhui Jia, Ming-Wei Chang, and William W. Cohen. 2023b. Subject-driven Text-to-Image Generation via Apprenticeship Learning. In Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023."},{"key":"e_1_2_2_9_1","volume-title":"Any Door: Zero-shot Object-level Image Customization. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR","author":"Chen Xi","year":"2024","unstructured":"Xi Chen, Lianghua Huang, Yu Liu, Yujun Shen, Deli Zhao, and Hengshuang Zhao. 2024. Any Door: Zero-shot Object-level Image Customization. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2024. 6593\u20136602."},{"key":"e_1_2_2_10_1","doi-asserted-by":"crossref","unstructured":"Wei Cheng Juncheng Mu Xianfang Zeng Xin Chen Anqi Pang Chi Zhang Zhibin Wang Bin Fu Gang Yu Ziwei Liu et al. 2024. MVPaint: Synchronized Multi-View Diffusion for Painting Anything 3D. arXiv preprint arXiv:2411.02336 (2024).","DOI":"10.1109\/CVPR52734.2025.00063"},{"key":"e_1_2_2_11_1","volume-title":"Improving Diffusion Models for Authentic Virtual Try-on in the Wild. arXiv preprint arXiv:2403.05139","author":"Choi Yisol","year":"2024","unstructured":"Yisol Choi, Sangkyung Kwak, Kyungmin Lee, Hyungwon Choi, and Jinwoo Shin. 2024. Improving Diffusion Models for Authentic Virtual Try-on in the Wild. arXiv preprint arXiv:2403.05139 (2024)."},{"key":"e_1_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.261"},{"key":"e_1_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01263"},{"key":"e_1_2_2_14_1","unstructured":"Blender Foundation. 2024. Blender. https:\/\/www.blender.org\/"},{"key":"e_1_2_2_15_1","unstructured":"Foundry. 2024. Mari. https:\/\/www.foundry.com\/products\/mari"},{"key":"e_1_2_2_16_1","volume-title":"The Eleventh International Conference on Learning Representations, ICLR","author":"Gal Rinon","year":"2023","unstructured":"Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H. Bermano, Gal Chechik, and Daniel Cohen-Or. 2023a. An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion. In The Eleventh International Conference on Learning Representations, ICLR 2023."},{"key":"e_1_2_2_17_1","volume-title":"Designing an encoder for fast personalization of text-to-image models. arXiv preprint arXiv:2302.12228","author":"Gal Rinon","year":"2023","unstructured":"Rinon Gal, Moab Arar, Yuval Atzmon, Amit H Bermano, Gal Chechik, and Daniel Cohen-Or. 2023b. Designing an encoder for fast personalization of text-to-image models. arXiv preprint arXiv:2302.12228 (2023)."},{"key":"e_1_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00442"},{"key":"e_1_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00995"},{"key":"e_1_2_2_20_1","volume-title":"Annual Conference on Neural Information Processing Systems 2017","author":"Heusel Martin","year":"2017","unstructured":"Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. 2017. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. In Annual Conference on Neural Information Processing Systems 2017, NeurIPS 2017."},{"key":"e_1_2_2_21_1","volume-title":"Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685","author":"Hu Edward J","year":"2021","unstructured":"Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2021. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)."},{"key":"e_1_2_2_22_1","volume-title":"IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR","author":"Hu Li","year":"2024","unstructured":"Li Hu. 2024. Animate anyone: Consistent and controllable image-to-video synthesis for character animation. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2024. 8153\u20138163."},{"key":"e_1_2_2_23_1","volume-title":"Material Anything: Generating Materials for Any 3D Object via Diffusion. arXiv preprint arXiv:2411.15138","author":"Huang Xin","year":"2024","unstructured":"Xin Huang, Tengfei Wang, Ziwei Liu, and Qing Wang. 2024b. Material Anything: Generating Materials for Any 3D Object via Diffusion. arXiv preprint arXiv:2411.15138 (2024)."},{"key":"e_1_2_2_24_1","volume-title":"From Parts to Whole: A Unified Reference Framework for Controllable Human Image Generation. arXiv preprint arXiv:2404.15267","author":"Huang Zehuan","year":"2024","unstructured":"Zehuan Huang, Hongxing Fan, Lipeng Wang, and Lu Sheng. 2024a. From Parts to Whole: A Unified Reference Framework for Controllable Human Image Generation. arXiv preprint arXiv:2404.15267 (2024)."},{"key":"e_1_2_2_25_1","volume-title":"Yandong Li, Han Zhang, Boqing Gong, Tingbo Hou, Huisheng Wang, and Yu-Chuan Su.","author":"Jia Xuhui","year":"2023","unstructured":"Xuhui Jia, Yang Zhao, Kelvin CK Chan, Yandong Li, Han Zhang, Boqing Gong, Tingbo Hou, Huisheng Wang, and Yu-Chuan Su. 2023. Taming encoder for zero fine-tuning image customization with text-to-image diffusion models. arXiv preprint arXiv:2304.02642 (2023)."},{"key":"e_1_2_2_26_1","volume-title":"Flexitex: Enhancing texture generation with visual guidance. arXiv preprint arXiv:2409.12431","author":"Jiang DaDong","year":"2024","unstructured":"DaDong Jiang, Xianghui Yang, Zibo Zhao, Sheng Zhang, Jiaao Yu, Zeqiang Lai, Shaoxiong Yang, Chunchao Guo, Xiaobo Zhou, and Zhihui Ke. 2024. Flexitex: Enhancing texture generation with visual guidance. arXiv preprint arXiv:2409.12431 (2024)."},{"key":"e_1_2_2_27_1","volume-title":"Text-Guided Texturing by Synchronized Multi-View Diffusion. arXiv preprint arXiv:2311.12891","author":"Liu Yuxin","year":"2023","unstructured":"Yuxin Liu, Minshan Xie, Hanyuan Liu, and Tien-Tsin Wong. 2023. Text-Guided Texturing by Synchronized Multi-View Diffusion. arXiv preprint arXiv:2311.12891 (2023)."},{"key":"e_1_2_2_28_1","volume-title":"Stable, Consistent and High-Quality Text-to-Texture Generation. arXiv preprint arXiv:2409.18401","author":"Lu Jiawei","year":"2024","unstructured":"Jiawei Lu, Yingpeng Zhang, Zengjun Zhao, He Wang, Kun Zhou, and Tianjia Shao. 2024. GenesisTex2: Stable, Consistent and High-Quality Text-to-Texture Generation. arXiv preprint arXiv:2409.18401 (2024)."},{"key":"e_1_2_2_29_1","volume-title":"Texture Atlas Compression Based on Repeated Content Removal. In SIGGRAPH Asia 2023 Conference Papers, SA 2023. 52:1\u201352:11","author":"Luo Yuzhe","year":"2023","unstructured":"Yuzhe Luo, Xiaogang Jin, Zherong Pan, Kui Wu, Qilong Kou, Xiajun Yang, and Xifeng Gao. 2023. Texture Atlas Compression Based on Repeated Content Removal. In SIGGRAPH Asia 2023 Conference Papers, SA 2023. 52:1\u201352:11."},{"key":"e_1_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01218"},{"key":"e_1_2_2_31_1","volume-title":"The Twelfth International Conference on Learning Representations, ICLR","author":"Pan Zijie","year":"2024","unstructured":"Zijie Pan, Jiachen Lu, Xiatian Zhu, and Li Zhang. 2024. Enhancing High-Resolution 3D Generation through Pixel-wise Gradient Clipping. In The Twelfth International Conference on Learning Representations, ICLR 2024."},{"key":"e_1_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3658217"},{"key":"e_1_2_2_33_1","article-title":"EASI-Tex: Edge-Aware Mesh Texturing from Single Image","volume":"43","author":"Kishore Perla Sai Raj","year":"2024","unstructured":"Sai Raj Kishore Perla, Yizhi Wang, Ali Mahdavi-Amiri, and Hao Zhang. 2024. EASI-Tex: Edge-Aware Mesh Texturing from Single Image. ACM Trans. Graph. 43, 4 (2024), 40:1\u201340:11.","journal-title":"ACM Trans. Graph."},{"key":"e_1_2_2_34_1","volume-title":"Sdxl: Improving latent diffusion models for high-resolution image synthesis. arXiv preprint arXiv:2307.01952","author":"Podell Dustin","year":"2023","unstructured":"Dustin Podell, Zion English, Kyle Lacey, Andreas Blattmann, Tim Dockhorn, Jonas M\u00fcller, Joe Penna, and Robin Rombach. 2023. Sdxl: Improving latent diffusion models for high-resolution image synthesis. arXiv preprint arXiv:2307.01952 (2023)."},{"key":"e_1_2_2_35_1","volume-title":"International Conference on Learning Representations.","author":"Poole Ben","year":"2022","unstructured":"Ben Poole, Ajay Jain, Jonathan T. Barron, and Ben Mildenhall. 2022. DreamFusion: Text-to-3D using 2D Diffusion. In International Conference on Learning Representations."},{"key":"e_1_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3588432.3591503"},{"key":"e_1_2_2_37_1","volume-title":"High-Resolution Image Synthesis with Latent Diffusion Models. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR","author":"Rombach Robin","year":"2022","unstructured":"Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Bj\u00f6rn Ommer. 2022. High-Resolution Image Synthesis with Latent Diffusion Models. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022. 10674\u201310685."},{"key":"e_1_2_2_38_1","volume-title":"DreamBooth: Fine Tuning Text-to-Image Diffusion Models for Subject-Driven Generation. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR","author":"Ruiz Nataniel","year":"2023","unstructured":"Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Yael Pritch, Michael Rubinstein, and Kfir Aberman. 2023. DreamBooth: Fine Tuning Text-to-Image Diffusion Models for Subject-Driven Generation. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023. 22500\u201322510."},{"key":"e_1_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00816"},{"key":"e_1_2_2_40_1","volume-title":"a single image to consistent multi-view diffusion base model. arXiv preprint arXiv:2310.15110","author":"Shi Ruoxi","year":"2023","unstructured":"Ruoxi Shi, Hansheng Chen, Zhuoyang Zhang, Minghua Liu, Chao Xu, Xinyue Wei, Linghao Chen, Chong Zeng, and Hao Su. 2023. Zero123++: a single image to consistent multi-view diffusion base model. arXiv preprint arXiv:2310.15110 (2023)."},{"key":"e_1_2_2_41_1","volume-title":"EMO: Emote Portrait Alive Generating Expressive Portrait Videos with Audio2Video Diffusion Model Under Weak Conditions. In Computer Vision - ECCV 2024 - 18th European Conference. 244\u2013260.","author":"Tian Linrui","year":"2024","unstructured":"Linrui Tian, Qi Wang, Bang Zhang, and Liefeng Bo. 2024. EMO: Emote Portrait Alive Generating Expressive Portrait Videos with Audio2Video Diffusion Model Under Weak Conditions. In Computer Vision - ECCV 2024 - 18th European Conference. 244\u2013260."},{"key":"e_1_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICME57554.2024.10687386"},{"key":"e_1_2_2_43_1","unstructured":"Qixun Wang Xu Bai Haofan Wang Zekui Qin and Anthony Chen. 2024a. InstantID: Zero-shot Identity-Preserving Generation in Seconds. arXiv preprint arXiv:2401.07519 (2024)."},{"key":"e_1_2_2_44_1","volume-title":"StableGarment: Garment-Centric Generation via Stable Diffusion. arXiv preprint arXiv:2403.10783","author":"Wang Rui","year":"2024","unstructured":"Rui Wang, Hailong Guo, Jiaming Liu, Huaxia Li, Haibo Zhao, Xu Tang, Yao Hu, Hao Tang, and Peipei Li. 2024c. StableGarment: Garment-Centric Generation via Stable Diffusion. arXiv preprint arXiv:2403.10783 (2024)."},{"key":"e_1_2_2_45_1","volume-title":"Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data. In International Conference on Computer Vision Workshops (ICCVW).","author":"Wang Xintao","year":"2021","unstructured":"Xintao Wang, Liangbin Xie, Chao Dong, and Ying Shan. 2021. Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data. In International Conference on Computer Vision Workshops (ICCVW)."},{"key":"e_1_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01461"},{"key":"e_1_2_2_47_1","volume-title":"TexRO: Generating Delicate Textures of 3D Models by Recursive Optimization. arXiv preprint arXiv:2403.15009","author":"Wu Jinbo","year":"2024","unstructured":"Jinbo Wu, Xing Liu, Chenming Wu, Xiaobo Gao, Jialun Liu, Xinqi Liu, Chen Zhao, Haocheng Feng, Errui Ding, and Jingdong Wang. 2024. TexRO: Generating Delicate Textures of 3D Models by Recursive Optimization. arXiv preprint arXiv:2403.15009 (2024)."},{"key":"e_1_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00701"},{"key":"e_1_2_2_49_1","volume-title":"Yuchen Fan, Omri Armstrong, Forrest Iandola, Yilei Li, Ita Lifshitz, and Rakesh Ranjan.","author":"Xiang Xiaoyu","year":"2024","unstructured":"Xiaoyu Xiang, Liat Sless Gorelik, Yuchen Fan, Omri Armstrong, Forrest Iandola, Yilei Li, Ita Lifshitz, and Rakesh Ranjan. 2024. Make-A-Texture: Fast Shape-Aware Texture Generation in 3 Seconds. arXiv preprint arXiv:2412.07766 (2024)."},{"key":"e_1_2_2_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3687931"},{"key":"e_1_2_2_51_1","volume-title":"Hallo: Hierarchical Audio-Driven Visual Synthesis for Portrait Image Animation. arXiv preprint arXiv:2406.08801","author":"Xu Mingwang","year":"2024","unstructured":"Mingwang Xu, Hui Li, Qingkun Su, Hanlin Shang, Liwei Zhang, Ce Liu, Jingdong Wang, Yao Yao, and Siyu zhu. 2024. Hallo: Hierarchical Audio-Driven Visual Synthesis for Portrait Image Animation. arXiv preprint arXiv:2406.08801 (2024)."},{"key":"e_1_2_2_52_1","volume-title":"IP-Adapter: Text Compatible Image Prompt Adapter for Text-to-Image Diffusion Models. arXiv preprint arxiv:2308.06721","author":"Ye Hu","year":"2023","unstructured":"Hu Ye, Jun Zhang, Sibo Liu, Xiao Han, and Wei Yang. 2023. IP-Adapter: Text Compatible Image Prompt Adapter for Text-to-Image Diffusion Models. arXiv preprint arxiv:2308.06721 (2023)."},{"key":"e_1_2_2_53_1","unstructured":"Yu-Ying Yeh Jia-Bin Huang Changil Kim Lei Xiao Thu Nguyen-Phuoc Numair Khan Cheng Zhang Manmohan Chandraker Carl S Marshall Zhao Dong et al. 2024. TextureDreamer: Image-guided Texture Synthesis through Geometry-aware Diffusion. arXiv preprint arXiv:2401.09416 (2024)."},{"key":"e_1_2_2_54_1","volume-title":"Paint-It: Text-to-Texture Synthesis via Deep Convolutional Texture Map Optimization and Physically-Based Rendering. arXiv preprint arXiv:2312.11360","author":"Youwang Kim","year":"2023","unstructured":"Kim Youwang, Tae-Hyun Oh, and Gerard Pons-Moll. 2023. Paint-It: Text-to-Texture Synthesis via Deep Convolutional Texture Map Optimization and Physically-Based Rendering. arXiv preprint arXiv:2312.11360 (2023)."},{"key":"e_1_2_2_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/3687909"},{"key":"e_1_2_2_56_1","volume-title":"Paint3D: Paint Anything 3D with Lighting-Less Texture Diffusion Models. arXiv preprint arXiv:2312.13913","author":"Zeng Xianfang","year":"2023","unstructured":"Xianfang Zeng, Xin Chen, Zhongqi Qi, Wen Liu, Zibo Zhao, Zhibin Wang, Bin Fu, Yong Liu, and Gang Yu. 2023. Paint3D: Paint Anything 3D with Lighting-Less Texture Diffusion Models. arXiv preprint arXiv:2312.13913 (2023)."},{"key":"e_1_2_2_57_1","volume-title":"ACM SIGGRAPH 2024 Conference Papers, SIGGRAPH 2024.","author":"Zhang Hongkun","year":"2024","unstructured":"Hongkun Zhang, Zherong Pan, Congyi Zhang, Lifeng Zhu, and Xifeng Gao. 2024b. Tex-painter: Generative mesh texturing with multi-view consistency. In ACM SIGGRAPH 2024 Conference Papers, SIGGRAPH 2024."},{"key":"e_1_2_2_58_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"e_1_2_2_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/3658146"},{"key":"e_1_2_2_60_1","volume-title":"FlashFace: Human Image Personalization with High-fidelity Identity Preservation. arXiv preprint arXiv:2403.17008","author":"Zhang Shilong","year":"2024","unstructured":"Shilong Zhang, Lianghua Huang, Xi Chen, Yifei Zhang, Zhi-Fan Wu, Yutong Feng, Wei Wang, Yujun Shen, Yu Liu, and Ping Luo. 2024a. FlashFace: Human Image Personalization with High-fidelity Identity Preservation. arXiv preprint arXiv:2403.17008 (2024)."}],"container-title":["ACM Transactions on Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3731158","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T17:56:02Z","timestamp":1774634162000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3731158"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,27]]},"references-count":60,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,8,1]]}},"alternative-id":["10.1145\/3731158"],"URL":"https:\/\/doi.org\/10.1145\/3731158","relation":{},"ISSN":["0730-0301","1557-7368"],"issn-type":[{"value":"0730-0301","type":"print"},{"value":"1557-7368","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,7,27]]},"assertion":[{"value":"2025-07-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}