{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T16:56:34Z","timestamp":1772643394874,"version":"3.50.1"},"reference-count":61,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s00371-025-04245-5","type":"journal-article","created":{"date-parts":[[2026,1,19]],"date-time":"2026-01-19T14:33:18Z","timestamp":1768833198000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DeepSketch2Wear: democratizing 3D garment creation via freehand sketches and text"],"prefix":"10.1007","volume":"42","author":[{"given":"Xinyu","family":"Chen","sequence":"first","affiliation":[]},{"given":"Jianqi","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Chaotao","family":"Ding","sequence":"additional","affiliation":[]},{"given":"Runlong","family":"Cao","sequence":"additional","affiliation":[]},{"given":"Lanyun","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Wenjun","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Ying","family":"Zang","sequence":"additional","affiliation":[]},{"given":"Tianrun","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,19]]},"reference":[{"key":"4245_CR1","volume-title":"Fashion Design: The Complete Guide","author":"J Hopkins","year":"2021","unstructured":"Hopkins, J.: Fashion Design: The Complete Guide. Bloomsbury Academic, London (2021)"},{"key":"4245_CR2","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1007\/s12186-018-09216-7","volume":"12","author":"V Caruso","year":"2019","unstructured":"Caruso, V., Cattaneo, A., Gurtner, J.-L., Ainsworth, S.: Professional vision in fashion design: practices and views of teachers and learners. Vocat. Learn. 12, 47\u201365 (2019)","journal-title":"Vocat. Learn."},{"key":"4245_CR3","doi-asserted-by":"crossref","unstructured":"Koma, K.: Fashion and identity in virtual spaces: the other bodies as an avatar in animal crossing. In: Sustainability Challenges in the Fashion Industry: Civilization Crisis. Decolonization, Cultural Legacy, and Transitions, pp. 187\u2013202. Springer, Cham, Switzerland (2023)","DOI":"10.1007\/978-981-99-0349-8_12"},{"key":"4245_CR4","doi-asserted-by":"crossref","unstructured":"Stergiou, M., Vosinakis, S.: Exploring costume-avatar interaction in digital dance experiences. In: Proceedings of the 8th International Conference on Movement and Computing, pp. 1\u20136 (2022)","DOI":"10.1145\/3537972.3537980"},{"key":"4245_CR5","doi-asserted-by":"crossref","unstructured":"Liao, C.: Virtual fashion play as embodied identity re\/assembling: second life fashion bloggers and their avatar bodies (pp. 101\u2013127) (2011)","DOI":"10.1007\/978-0-85729-361-9_6"},{"key":"4245_CR6","unstructured":"Tepe, J.: Wearing digital bodies: designing and experiencing dress as poly-body objects at the intersection of the physical and the digital. In: 24th IFFTI Conference, Nottingham, England, 5th-8th April, 2022., pp. 98\u2013315 (2022)"},{"issue":"3","key":"4245_CR7","doi-asserted-by":"publisher","first-page":"279","DOI":"10.1386\/jgvw.8.3.279_1","volume":"8","author":"B Koles","year":"2016","unstructured":"Koles, B., Nagy, P.: Avatars as transitional objects: the impact of avatars and digital objects on adolescent gamers. J. Gam. Virtual Worlds 8(3), 279\u2013296 (2016)","journal-title":"J. Gam. Virtual Worlds"},{"key":"4245_CR8","doi-asserted-by":"crossref","unstructured":"Bandyopadhyay, H., Koley, S., Das, A., Bhunia, A.K., Sain, A., Chowdhury, P.N., Xiang, T., Song, Y.-Z.: Doodle your 3d: From abstract freehand sketches to precise 3d shapes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9795\u20139805 (2024)","DOI":"10.1109\/CVPR52733.2024.00935"},{"key":"4245_CR9","doi-asserted-by":"crossref","unstructured":"Kwan, K.C., Fu, H.: Mobi3dsketch: 3d sketching in mobile ar. In: CHI 2019, pp. 1\u201311 (2019)","DOI":"10.1145\/3290605.3300406"},{"key":"4245_CR10","doi-asserted-by":"crossref","unstructured":"Arora, R., Kazi, R.H., Anderson, F., Grossman, T., Singh, K., Fitzmaurice, G.W.: Experimental evaluation of sketching on surfaces in vr. In: CHI 2017, vol. 17, pp. 5643\u20135654 (2017)","DOI":"10.1145\/3025453.3025474"},{"key":"4245_CR11","doi-asserted-by":"crossref","unstructured":"Machuca, M.D.B., Stuerzlinger, W., Asente, P.: Smart3dguides: making unconstrained immersive 3d drawing more accurate. In: VRST, pp. 1\u201313 (2019)","DOI":"10.1145\/3359996.3364254"},{"key":"4245_CR12","doi-asserted-by":"crossref","unstructured":"Zhang, S.-H., Guo, Y.C., Gu, Q.-W.: Sketch2model: view-aware 3d modeling from single free-hand sketches. Cornell University - arXiv, Cornell University - arXiv (2021)","DOI":"10.1109\/CVPR46437.2021.00595"},{"key":"4245_CR13","doi-asserted-by":"crossref","unstructured":"Chen, T., Fu, C., Zhu, L., Mao, P., Zhang, J., Zang, Y., Sun, L.: Deep3dsketch: 3d modeling from free-hand sketches with view-and structural-aware adversarial training. In: ICASSP, pp. 1\u20135. IEEE (2023)","DOI":"10.1109\/ICASSP49357.2023.10096348"},{"key":"4245_CR14","doi-asserted-by":"crossref","unstructured":"Chen, T., Fu, C., Zang, Y., Zhu, L., Zhang, J., Mao, P., Sun, L.: Deep3dsketch+: Rapid 3d modeling from single free-hand sketches. In: International Conference on Multimedia Modeling, pp. 16\u201328. Springer (2023)","DOI":"10.1007\/978-3-031-27818-1_2"},{"key":"4245_CR15","doi-asserted-by":"crossref","unstructured":"Zang, Y., Ding, C., Chen, T., Mao, P., Hu, W.: Deep3dsketch+$$\\backslash $$+: High-fidelity 3d modeling from single free-hand sketches. In: 2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 1537\u20131542. IEEE (2023)","DOI":"10.1109\/SMC53992.2023.10393936"},{"issue":"1","key":"4245_CR16","doi-asserted-by":"publisher","first-page":"149","DOI":"10.1631\/FITEE.2300314","volume":"25","author":"T Chen","year":"2024","unstructured":"Chen, T., Cao, R., Li, Z., Zang, Y., Sun, L.: Deep3dsketch-im: rapid high-fidelity ai 3d model generation by single freehand sketches. Front. Inf. Technol. Electron. Eng. 25(1), 149\u2013159 (2024)","journal-title":"Front. Inf. Technol. Electron. Eng."},{"key":"4245_CR17","doi-asserted-by":"crossref","unstructured":"Chen, T., Cao, R., Lu, A., Xu, T., Zhang, X., Papa, M., Zhang, M., Sun, L., Zang, Y.: High-fidelity 3d model generation with relightable appearance from single freehand sketches and text guidance. In: 2024 IEEE International Conference on Multimedia and Expo Workshops (ICMEW), pp. 1\u20136. IEEE (2024)","DOI":"10.1109\/ICMEW63481.2024.10645361"},{"key":"4245_CR18","doi-asserted-by":"crossref","unstructured":"Guillard, B., Remelli, E., Yvernay, P., Fua, P.: Sketch2mesh: Reconstructing and editing 3d shapes from sketches. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13023\u201313032 (2021)","DOI":"10.1109\/ICCV48922.2021.01278"},{"issue":"4","key":"4245_CR19","first-page":"1","volume":"42","author":"X-Y Zheng","year":"2023","unstructured":"Zheng, X.-Y., Pan, H., Wang, P.-S., Tong, X., Liu, Y., Shum, H.-Y.: Locally attentional sdf diffusion for controllable 3d shape generation. ACM Trans. Gr. (ToG) 42(4), 1\u201313 (2023)","journal-title":"ACM Trans. Gr. (ToG)"},{"key":"4245_CR20","doi-asserted-by":"crossref","unstructured":"Gao, C., Yu, Q., Sheng, L., Song, Y.-Z., Xu, D.: Sketchsampler: Sketch-based 3d reconstruction via view-dependent depth sampling. In: European Conference on Computer Vision, pp. 464\u2013479. Springer (2022)","DOI":"10.1007\/978-3-031-19769-7_27"},{"key":"4245_CR21","doi-asserted-by":"crossref","unstructured":"Chen, T., Ding, C., Zhang, S., Yu, C., Zang, Y., Li, Z., Peng, S., Sun, L.: Rapid 3d model generation with intuitive 3d input. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12554\u201312564 (2024)","DOI":"10.1109\/CVPR52733.2024.01193"},{"issue":"4","key":"4245_CR22","first-page":"370","volume":"33","author":"A Bonnici","year":"2019","unstructured":"Bonnici, A., Akman, A., Calleja, G., Camilleri, K.P., Fehling, P., Ferreira, A., Hermuth, F., Israel, J.H., Landwehr, T., Liu, J., et al.: Sketch-based interaction and modeling: Where do we stand? AI EDAM 33(4), 370\u2013388 (2019)","journal-title":"AI EDAM"},{"issue":"1","key":"4245_CR23","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1016\/j.cag.2008.09.013","volume":"33","author":"L Olsen","year":"2009","unstructured":"Olsen, L., Samavati, F.F., Sousa, M.C., Jorge, J.A.: Sketch-based modeling: A survey. Comput. Gr. 33(1), 85\u2013103 (2009)","journal-title":"Comput. Gr."},{"key":"4245_CR24","doi-asserted-by":"crossref","unstructured":"Chen, T., Ding, C., Zhu, L., Zang, Y., Liao, Y., Li, Z., Sun, L.: Reality3dsketch: rapid 3d modeling of objects from single freehand sketches. arXiv preprint arXiv:2310.18148 (2023)","DOI":"10.1109\/TMM.2023.3327533"},{"key":"4245_CR25","doi-asserted-by":"crossref","unstructured":"Alldieck, T., Pons-Moll, G., Theobalt, C., Magnor, M.: Tex2shape: Detailed full human body geometry from a single image. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2293\u20132303 (2019)","DOI":"10.1109\/ICCV.2019.00238"},{"key":"4245_CR26","doi-asserted-by":"crossref","unstructured":"Saito, S., Huang, Z., Natsume, R., Morishima, S., Kanazawa, A., Li, H.: Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2304\u20132314 (2019)","DOI":"10.1109\/ICCV.2019.00239"},{"key":"4245_CR27","doi-asserted-by":"crossref","unstructured":"Zhao, F., Wang, W., Liao, S., Shao, L.: Learning anchored unsigned distance functions with gradient direction alignment for single-view garment reconstruction. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 12674\u201312683 (2021)","DOI":"10.1109\/ICCV48922.2021.01244"},{"issue":"4","key":"4245_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530179","volume":"41","author":"M Korosteleva","year":"2022","unstructured":"Korosteleva, M., Lee, S.-H.: Neuraltailor: reconstructing sewing pattern structures from 3d point clouds of garments. ACM Trans. Gr. (TOG) 41(4), 1\u201316 (2022)","journal-title":"ACM Trans. Gr. (TOG)"},{"issue":"4","key":"4245_CR29","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3658147","volume":"43","author":"K He","year":"2024","unstructured":"He, K., Yao, K., Zhang, Q., Yu, J., Liu, L., Xu, L.: Dresscode: autoregressively sewing and generating garments from text guidance. ACM Trans. Gr. (TOG) 43(4), 1\u201313 (2024)","journal-title":"ACM Trans. Gr. (TOG)"},{"key":"4245_CR30","doi-asserted-by":"crossref","unstructured":"Sarafianos, N., Stuyck, T., Xiang, X., Li, Y., Popovic, J., Ranjan, R.: Garment3dgen: 3d garment stylization and texture generation. arXiv preprint arXiv:2403.18816 (2024)","DOI":"10.1109\/3DV66043.2025.00131"},{"key":"4245_CR31","doi-asserted-by":"crossref","unstructured":"Tatsukawa, Y., Qi, A., Shen, I.-C., Igarashi, T.: Garmentimage: Raster encoding of garment sewing patterns with diverse topologies. In: Proceedings of the Special Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers, pp. 1\u201311 (2025)","DOI":"10.1145\/3721238.3730632"},{"issue":"3\u20134","key":"4245_CR32","doi-asserted-by":"publisher","first-page":"1889","DOI":"10.1002\/cav.1889","volume":"30","author":"Z Hu","year":"2019","unstructured":"Hu, Z., Xie, H., Fukusato, T., Sato, T., Igarashi, T.: Sketch2vf: sketch-based flow design with conditional generative adversarial network. Comput. Anim. Virt. Worlds 30(3\u20134), 1889 (2019)","journal-title":"Comput. Anim. Virt. Worlds"},{"key":"4245_CR33","doi-asserted-by":"crossref","unstructured":"He, Y., Xie, H., Zhang, C., Yang, X., Miyata, K.: Sketch-based normal map generation with geometric sampling. In: International Workshop on Advanced Imaging Technology (IWAIT) 2021, vol. 11766, pp. 261\u2013266. SPIE (2021)","DOI":"10.1117\/12.2590760"},{"key":"4245_CR34","doi-asserted-by":"crossref","unstructured":"Xie, H., Arihara, K., Sato, S., Miyata, K.: Dualsmoke: Sketch-based smoke illustration design with two-stage generative model. Comput. Vis. Media, 1\u201315 (2024)","DOI":"10.1007\/s41095-022-0318-0"},{"issue":"4","key":"4245_CR35","doi-asserted-by":"publisher","first-page":"459","DOI":"10.1587\/transinf.2022IIP0011","volume":"106","author":"Y Peng","year":"2023","unstructured":"Peng, Y., Zhao, C., Xie, H., Fukusato, T., Miyata, K., Igarashi, T.: Dualmotion: global-to-local casual motion design for character animations. IEICE Trans. Inf. Syst. 106(4), 459\u2013468 (2023)","journal-title":"IEICE Trans. Inf. Syst."},{"key":"4245_CR36","doi-asserted-by":"publisher","first-page":"63","DOI":"10.1007\/s41095-021-0227-7","volume":"8","author":"Z Huang","year":"2022","unstructured":"Huang, Z., Peng, Y., Hibino, T., Zhao, C., Xie, H., Fukusato, T., Miyata, K.: dualface: two-stage drawing guidance for freehand portrait sketching. Comput. Vis. Media 8, 63\u201377 (2022)","journal-title":"Comput. Vis. Media"},{"issue":"2","key":"4245_CR37","doi-asserted-by":"publisher","first-page":"1367","DOI":"10.1007\/s00371-024-03425-z","volume":"41","author":"Y Hu","year":"2025","unstructured":"Hu, Y., Zhang, J., Bai, L., Li, J., Li, B., Zang, Y., Hu, W.: From sketch to reality: precision-friendly 3d generation technology. Vis. Comput. 41(2), 1367\u20131378 (2025)","journal-title":"Vis. Comput."},{"key":"4245_CR38","doi-asserted-by":"crossref","unstructured":"Wang, T.Y., Ceylan, D., Popovic, J., Mitra, N.J.: Learning a shared shape space for multimodal garment design. arXiv preprint arXiv:1806.11335 (2018)","DOI":"10.1145\/3272127.3275074"},{"key":"4245_CR39","doi-asserted-by":"crossref","unstructured":"Chen, Y., Xian, C., Jin, S., Li, G.: 3d shape-adapted garment generation with sketches. In: Advances in Computer Graphics: 38th Computer Graphics International Conference, CGI 2021, Virtual Event, September 6\u201310, 2021, Proceedings 38, pp. 125\u2013136. Springer (2021)","DOI":"10.1007\/978-3-030-89029-2_10"},{"key":"4245_CR40","doi-asserted-by":"crossref","unstructured":"He, Y., Xie, H., Miyata, K.: Sketch2cloth: Sketch-based 3d garment generation with unsigned distance fields. In: 2023 Nicograph International (NicoInt), pp. 38\u201345. IEEE (2023)","DOI":"10.1109\/NICOINT59725.2023.00016"},{"issue":"1","key":"4245_CR41","first-page":"43","volume":"6","author":"M Jiang","year":"2024","unstructured":"Jiang, M., Tian, Z., Yu, C., Shi, Y., Liu, L., Peng, T., Hu, X., Yu, F.: Intelligent 3d garment system of the human body based on deep spiking neural network. Virt. Real. Intell. Hardw. 6(1), 43\u201355 (2024)","journal-title":"Virt. Real. Intell. Hardw."},{"issue":"4","key":"4245_CR42","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1145\/37402.37422","volume":"21","author":"WE Lorensen","year":"1987","unstructured":"Lorensen, W.E., Cline, H.E.: Marching cubes: a high resolution 3d surface construction algorithm. ACM siggr. Comput. Gr. 21(4), 163\u2013169 (1987)","journal-title":"ACM siggr. Comput. Gr."},{"key":"4245_CR43","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Adv. Neural. Inf. Process. Syst. 33, 6840\u20136851 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4245_CR44","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"4245_CR45","doi-asserted-by":"publisher","first-page":"528","DOI":"10.1016\/j.neucom.2020.05.124","volume":"456","author":"Y Zheng","year":"2021","unstructured":"Zheng, Y., Yao, H., Sun, X., Zhang, S., Zhao, S., Porikli, F.: Sketch-specific data augmentation for freehand sketch recognition. Neurocomputing 456, 528\u2013539 (2021)","journal-title":"Neurocomputing"},{"key":"4245_CR46","doi-asserted-by":"crossref","unstructured":"Shaheen, S., Affara, L., Ghanem, B.: Constrained convolutional sparse coding for parametric based reconstruction of line drawings, 4424\u20134432 (2017)","DOI":"10.1109\/ICCV.2017.474"},{"key":"4245_CR47","doi-asserted-by":"crossref","unstructured":"Liu, Y., Xie, M., Liu, H., Wong, T.-T.: Text-guided texturing by synchronized multi-view diffusion. In: SIGGRAPH Asia 2024 Conference Papers, pp. 1\u201311 (2024)","DOI":"10.1145\/3680528.3687621"},{"key":"4245_CR48","doi-asserted-by":"crossref","unstructured":"Zhu, H., Cao, Y., Jin, H., Chen, W., Du, D., Wang, Z., Cui, S., Han, X.: Deep fashion3d: a dataset and benchmark for 3d garment reconstruction from single images. In: European Conference on Computer Vision, pp. 512\u2013530. Springer (2020)","DOI":"10.1007\/978-3-030-58452-8_30"},{"key":"4245_CR49","first-page":"679","volume":"6","author":"J Canny","year":"2009","unstructured":"Canny, J.: A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 6, 679\u2013698 (2009)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4245_CR50","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"key":"4245_CR51","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"4245_CR52","doi-asserted-by":"crossref","unstructured":"Zhang, L., Rao, A., Agrawala, M.: Adding conditional control to text-to-image diffusion models. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3836\u20133847 (2023)","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"4245_CR53","doi-asserted-by":"crossref","unstructured":"Cheng, Y.-C., Lee, H.-Y., Tulyakov, S., Schwing, A., Gui, L.: Sdfusion: multimodal 3d shape completion, reconstruction, and generation (2022)","DOI":"10.1109\/CVPR52729.2023.00433"},{"key":"4245_CR54","first-page":"31841","volume":"35","author":"J Gao","year":"2022","unstructured":"Gao, J., Shen, T., Wang, Z., Chen, W., Yin, K., Li, D., Litany, O., Gojcic, Z., Fidler, S.: Get3d: a generative model of high quality 3d textured shapes learned from images. Adv. Neural. Inf. Process. Syst. 35, 31841\u201331854 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4245_CR55","doi-asserted-by":"crossref","unstructured":"Sanghi, A., Fu, R., Liu, V., Willis, K.D., Shayani, H., Khasahmadi, A.H., Sridhar, S., Ritchie, D.: Clip-sculptor: Zero-shot generation of high-fidelity and diverse shapes from natural language. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18339\u201318348 (2023)","DOI":"10.1109\/CVPR52729.2023.01759"},{"key":"4245_CR56","doi-asserted-by":"crossref","unstructured":"Cai, Y., Wang, Y., Zhu, Y., Cham, T.-J., Cai, J., Yuan, J., Liu, J., Zheng, C., Yan, S., Ding, H., et al.: A unified 3d human motion synthesis model via conditional variational auto-encoder. In: ICCV, pp. 11645\u201311655 (2021)","DOI":"10.1109\/ICCV48922.2021.01144"},{"key":"4245_CR57","doi-asserted-by":"crossref","unstructured":"Michel, O., Bar-On, R., Liu, R., Benaim, S., Hanocka, R.: Text2mesh: text-driven neural stylization for meshes. In: CVPR, pp. 13492\u201313502 (2022)","DOI":"10.1109\/CVPR52688.2022.01313"},{"key":"4245_CR58","unstructured":"Yao, S., Zhong, R., Yan, Y., Zhai, G., Yang, X.: Dfa-nerf: Personalized talking head generation via disentangled face attributes neural rendering. arXiv preprint arXiv:2201.00791 (2022)"},{"key":"4245_CR59","volume-title":"Measuring the User Experience: Collecting, Analyzing, and Presenting UX Metrics","author":"B Albert","year":"2022","unstructured":"Albert, B., Tullis, T.: Measuring the User Experience: Collecting, Analyzing, and Presenting UX Metrics. Morgan Kaufmann, Burlington, MA, USA (2022)"},{"key":"4245_CR60","doi-asserted-by":"crossref","unstructured":"Oh, C., Song, J., Choi, J., Kim, S., Lee, S., Suh, B.: I lead, you help but only with enough details: understanding user experience of co-creation with artificial intelligence. In: CHI 2018, pp. 1\u201313 (2018)","DOI":"10.1145\/3173574.3174223"},{"key":"4245_CR61","doi-asserted-by":"crossref","unstructured":"Yu, Z., Dou, Z., Long, X., Lin, C., Li, Z., Liu, Y., M\u00fcller, N., Komura, T., Habermann, M., Theobalt, C., et al.: Surf-d: Generating high-quality surfaces of arbitrary topologies using diffusion models. In: European Conference on Computer Vision, pp. 419\u2013438. Springer (2025)","DOI":"10.1007\/978-3-031-72933-1_24"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04245-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04245-5","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04245-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T12:45:35Z","timestamp":1772628335000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04245-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1]]},"references-count":61,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["4245"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04245-5","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1]]},"assertion":[{"value":"11 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 October 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}],"article-number":"130"}}