{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T13:50:58Z","timestamp":1765547458581,"version":"build-2065373602"},"reference-count":44,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T00:00:00Z","timestamp":1757289600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T00:00:00Z","timestamp":1757289600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100003790","name":"Hiroshima University","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100003790","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Machine Vision and Applications"],"published-print":{"date-parts":[[2025,11]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>\n                    Despite the significant potential of 3D object editing to impact various industries, recent research in 3D generation and editing has primarily focused on converting text and images into 3D models, often paying limited attention to the need for\n                    <jats:italic>fine-grained control<\/jats:italic>\n                    over existing 3D objects. This paper introduces a framework that uses a pre-trained regressor to enable continuous and attribute-specific modifications of both the stylistic and geometric attributes of 3D vehicle models. Here, \u201cfine-grained control\u201d refers to the ability to adjust specific geometric or stylistic attributes (such as roof length or perceived luxury) in a continuous and independent manner. Our method aims to preserve the identity of vehicle 3D objects and support multi-attribute editing, allowing for extensive customization while maintaining the model\u2019s structural integrity. The framework leverages DeepSDF to obtain latent representations suitable for continuous attribute editing. Experimental results demonstrate the effectiveness of our approach in achieving detailed, controlled edits on a variety of vehicle 3D models. The code is released at\n                    <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" xlink:href=\"https:\/\/github.com\/JiangDong-miao\/Vehicle_LatentEdit\" ext-link-type=\"uri\">https:\/\/github.com\/JiangDong-miao\/Vehicle_LatentEdit<\/jats:ext-link>\n                    .\n                  <\/jats:p>","DOI":"10.1007\/s00138-025-01739-z","type":"journal-article","created":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T10:00:16Z","timestamp":1757325616000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Fine-grained 3D vehicle shape manipulation via latent space editing"],"prefix":"10.1007","volume":"36","author":[{"given":"JiangDong","family":"Miao","sequence":"first","affiliation":[]},{"given":"Tatsuya","family":"Ikeda","sequence":"additional","affiliation":[]},{"given":"Bisser","family":"Raytchev","sequence":"additional","affiliation":[]},{"given":"Ryota","family":"Mizoguchi","sequence":"additional","affiliation":[]},{"given":"Takenori","family":"Hiraoka","sequence":"additional","affiliation":[]},{"given":"Takuji","family":"Nakashima","sequence":"additional","affiliation":[]},{"given":"Keigo","family":"Shimizu","sequence":"additional","affiliation":[]},{"given":"Toru","family":"Higaki","sequence":"additional","affiliation":[]},{"given":"Kazufumi","family":"Kaneda","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,8]]},"reference":[{"key":"1739_CR1","doi-asserted-by":"publisher","unstructured":"Arechiga, N., Permenter, F., Song, B., Yuan, C.: Drag-guided diffusion models for vehicle image generation. In: NeurIPS, (2023). https:\/\/doi.org\/10.48550\/arXiv.2306.09935","DOI":"10.48550\/arXiv.2306.09935"},{"key":"1739_CR2","unstructured":"Morita, H., Shintani, K., Yuan, C., Permenter, F.: VehicleSDF: A 3D generative model for constrained engineering design via surrogate modeling (2024). arXiv:org\/abs\/2410.18986"},{"key":"1739_CR3","unstructured":"Elrefaie, M., Morar, F., Dai, A., Ahmed, F.: DrivAerNet++: A Large-Scale Multimodal Car Dataset with Computational Fluid Dynamics Simulations and Deep Learning Benchmarks (2025). arXiv:org\/abs\/2406.09624"},{"key":"1739_CR4","doi-asserted-by":"publisher","unstructured":"M\u00e4tt\u00e4, T., Sharifipour, S., L\u00f3pez, M.B., Casado, C.\u00c1.: Spatio-temporal 3d point clouds from wi-fi-csi data via transformer networks. In: 2025 IEEE 5th International Symposium on Joint Communications & Sensing (JC &S), pp. 1\u20136 (2025). https:\/\/doi.org\/10.1109\/JCS64661.2025.10880635","DOI":"10.1109\/JCS64661.2025.10880635"},{"key":"1739_CR5","doi-asserted-by":"publisher","unstructured":"Zhuang, J., Wang, C., Liu, L., Lin, L., Li, G.: Dreameditor: Text-driven 3d scene editing with neural fields. In: SIGGRAPH Asia, p. (2023). https:\/\/doi.org\/10.1145\/3610548.3618190","DOI":"10.1145\/3610548.3618190"},{"key":"1739_CR6","doi-asserted-by":"publisher","unstructured":"Li, Y., Dou, Y., Shi, Y., Lei, Y., Chen, X., Zhang, Y., Zhou, P., Ni, B.: FocalDreamer: Text-driven 3D Editing via Focal-fusion Assembly (2024). https:\/\/doi.org\/10.1609\/aaai.v38i4.28113","DOI":"10.1609\/aaai.v38i4.28113"},{"key":"1739_CR7","doi-asserted-by":"publisher","unstructured":"Tzathas, P., Maragos, P., Roussos, A.: 3d neural sculpting (3dns): Editing neural signed distance functions. In: WACV, pp. 4510\u20134519 (2023). https:\/\/doi.org\/10.1109\/WACV56688.2023.00450","DOI":"10.1109\/WACV56688.2023.00450"},{"key":"1739_CR8","doi-asserted-by":"publisher","unstructured":"Mikaeili, A., Perel, O., Safaee, M., Cohen-Or, D., Mahdavi-Amiri, A.: Sked: Sketch-guided text-based 3d editing. In: ICCV, pp. 14561\u201314573 (2023). https:\/\/doi.org\/10.1109\/ICCV51070.2023.01343","DOI":"10.1109\/ICCV51070.2023.01343"},{"key":"1739_CR9","doi-asserted-by":"publisher","unstructured":"Zhuang, J., Kang, D., Cao, Y.-P., Li, G., Lin, L., Shan, Y.: Tip-editor: An accurate 3d editor following both text-prompts and image-prompts. In: SIGGRAPH, p. (2024). https:\/\/doi.org\/10.48550\/arXiv.2401.14828","DOI":"10.48550\/arXiv.2401.14828"},{"key":"1739_CR10","doi-asserted-by":"publisher","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: Nerf: Representing scenes as neural radiance fields for view synthesis. In: ECCV, p. (2020). https:\/\/doi.org\/10.48550\/arXiv.2003.08934","DOI":"10.48550\/arXiv.2003.08934"},{"key":"1739_CR11","doi-asserted-by":"publisher","unstructured":"Martin-Brualla, R., Radwan, N., Sajjadi, M.S.M., Barron, J.T., Dosovitskiy, A., Duckworth, D.: Nerf in the wild: Neural radiance fields for unconstrained photo collections. In: CVPR, (2021). https:\/\/doi.org\/10.48550\/arXiv.2008.02268","DOI":"10.48550\/arXiv.2008.02268"},{"key":"1739_CR12","doi-asserted-by":"publisher","unstructured":"Barron, J.T., Mildenhall, B., Tancik, M., Hedman, P., Martin-Brualla, R., Srinivasan, P.P.: A multiscale representation for anti-aliasing neural radiance fields. In: ICCV, (2021). https:\/\/doi.org\/10.48550\/arXiv.2103.13415","DOI":"10.48550\/arXiv.2103.13415"},{"issue":"4","key":"1739_CR13","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3592433","volume":"42","author":"B Kerbl","year":"2023","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., Drettakis, G.: 3d gaussian splatting for real-time radiance field rendering. TOG 42(4), 139\u20131 (2023)","journal-title":"TOG"},{"key":"1739_CR14","doi-asserted-by":"crossref","unstructured":"Madhavaram, V., Rawat, S., Devaguptapu, C., Sharma, C., Kaul, M.: Towards a Training Free Approach for 3D Scene Editing (2024). arXiv:org\/abs\/2412.12766","DOI":"10.1109\/WACV61041.2025.00286"},{"issue":"7","key":"1739_CR15","doi-asserted-by":"publisher","first-page":"e15215","DOI":"10.1111\/cgf.15215","volume":"43","author":"Y Sun","year":"2024","unstructured":"Sun, Y., Tian, R., Han, X., Liu, X., Zhang, Y., Xu, K.: Gseditpro: 3d gaussian splatting editing with attention-based progressive localization. Comput. Grap. Forum 43(7), e15215 (2024). https:\/\/doi.org\/10.1111\/cgf.15215","journal-title":"Comput. Grap. Forum"},{"key":"1739_CR16","unstructured":"Nam, G., Khlifi, M., Rodriguez, A., Tono, A., Zhou, L., Guerrero, P.: 3D-LDM: Neural Implicit 3D Shape Generation with Latent Diffusion Models (2022). arXiv:org\/abs\/2212.00842"},{"key":"1739_CR17","doi-asserted-by":"crossref","unstructured":"Li, M., Duan, Y., Zhou, J., Lu, J.: Diffusion-SDF: Text-to-Shape via Voxelized Diffusion (2023). arXiv:org\/abs\/2212.03293","DOI":"10.1109\/CVPR52729.2023.01216"},{"key":"1739_CR18","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-Resolution Image Synthesis with Latent Diffusion Models (2022). arXiv:org\/abs\/2112.10752","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"1739_CR19","unstructured":"Poole, B., Jain, A., Barron, J.T., Mildenhall, B.: DreamFusion: Text-to-3D using 2D Diffusion (2022). arXiv:org\/abs\/2209.14988"},{"key":"1739_CR20","doi-asserted-by":"publisher","unstructured":"Park, J.J., Florence, P., Straub, J., Newcombe, R., Lovegrove, S.: Deepsdf: Learning continuous signed distance functions for shape representation. In: CVPR, pp. 165\u2013174 (2019). https:\/\/doi.org\/10.48550\/arXiv.1901.05103","DOI":"10.48550\/arXiv.1901.05103"},{"key":"1739_CR21","doi-asserted-by":"publisher","DOI":"10.1145\/37401.37422","author":"L William","year":"1987","unstructured":"William, L., Harvey, C.: Marching cubes: a high resolution 3d surface construction algorithm. ACM SIGGRAPH Comput. Graph. (1987). https:\/\/doi.org\/10.1145\/37401.37422","journal-title":"ACM SIGGRAPH Comput. Graph."},{"key":"1739_CR22","unstructured":"Goodfellow, I.J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. Adv. Neural Inform. Process. Syst. 27 (2014)"},{"key":"1739_CR23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00813","author":"T Karras","year":"2020","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. CVPR (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00813","journal-title":"CVPR"},{"key":"1739_CR24","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of gans for improved quality, stability, and variation. In: International Conference on Learning Representations, p. (2018). https:\/\/openreview.net\/forum?id=Hk99zCeAb"},{"key":"1739_CR25","unstructured":"Kingma, D.P., Welling, M.: Auto-Encoding Variational Bayes (2022). arXiv:org\/abs\/1312.6114"},{"key":"1739_CR26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00926","author":"Y Shen","year":"2020","unstructured":"Shen, Y., Gu, J., Tang, X., Zhou, B.: Interpreting the latent space of gans for semantic face editing. CVPR (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00926","journal-title":"CVPR"},{"key":"1739_CR27","doi-asserted-by":"publisher","unstructured":"H\u00e4rk\u00f6nen, E., Hertzmann, A., Lehtinen, J., Paris, S.: Ganspace: Discovering interpretable gan controls. In: NeurIPS, (2020). https:\/\/doi.org\/10.48550\/arXiv.2004.02546","DOI":"10.48550\/arXiv.2004.02546"},{"key":"1739_CR28","doi-asserted-by":"publisher","unstructured":"Voynov, A., Babenko, A.: Unsupervised discovery of interpretable directions in the gan latent space. In: ICML, (2020). https:\/\/doi.org\/10.48550\/arXiv.2002.03754","DOI":"10.48550\/arXiv.2002.03754"},{"issue":"12","key":"1739_CR29","doi-asserted-by":"publisher","first-page":"14590","DOI":"10.1109\/TPAMI.2023.3298868","volume":"45","author":"Y Liu","year":"2023","unstructured":"Liu, Y., Li, Q., Deng, Q., Sun, Z., Yang, M.-H.: Gan-based facial attribute manipulation. IEEE Trans. Pattern Anal. Mach. Intell. 45(12), 14590\u201314610 (2023). https:\/\/doi.org\/10.1109\/TPAMI.2023.3298868","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1739_CR30","doi-asserted-by":"crossref","unstructured":"Wu, Y., Thalmann, N.M., Thalmann, D.: A plastic-visco-elastic model for wrinkles in facial animation and skin aging. In: Proceedings of the Second Pacific Conference on Fundamentals of Computer Graphics, pp. 201\u2013213. World Scientific Publishing Co., Inc., USA (1995)","DOI":"10.1142\/9789814503860_0013"},{"issue":"5","key":"1739_CR31","doi-asserted-by":"publisher","first-page":"70","DOI":"10.1109\/38.403830","volume":"15","author":"DA Rowland","year":"1995","unstructured":"Rowland, D.A., Perrett, D.I.: Manipulating facial appearance through shape and color. IEEE Comput. Graphics Appl. 15(5), 70\u201376 (1995). https:\/\/doi.org\/10.1109\/38.403830","journal-title":"IEEE Comput. Graphics Appl."},{"issue":"8\u20139","key":"1739_CR32","doi-asserted-by":"publisher","first-page":"2002","DOI":"10.1007\/s11263-020-01308-z","volume":"128","author":"X Zheng","year":"2020","unstructured":"Zheng, X., Guo, Y., Huang, H., Li, Y., He, R.: A survey of deep facial attribute analysis. Int. J. Comput. Vision 128(8\u20139), 2002\u20132034 (2020). https:\/\/doi.org\/10.1007\/s11263-020-01308-z","journal-title":"Int. J. Comput. Vision"},{"key":"1739_CR33","unstructured":"Nickabadi, A., Fard, M.S., Farid, N.M., Mohammadbagheri, N.: A comprehensive survey on semantic facial attribute editing using generative adversarial networks. ArXiv arXiv:abs\/2205.10587 (2022)"},{"key":"1739_CR34","doi-asserted-by":"publisher","unstructured":"Fang, H., Deng, W., Zhong, Y., Hu, J.: Triple-gan: Progressive face aging with triple translation loss. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 3500\u20133509 (2020). https:\/\/doi.org\/10.1109\/CVPRW50498.2020.00410","DOI":"10.1109\/CVPRW50498.2020.00410"},{"key":"1739_CR35","doi-asserted-by":"publisher","unstructured":"Lee, Y.-H., Lai, S.-H.: Byeglassesgan: Identity preserving eyeglasses removal for face images. In: Computer Vision \u2013 ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXIX, pp. 243\u2013258. Springer, Berlin, Heidelberg (2020). https:\/\/doi.org\/10.1007\/978-3-030-58526-6_15","DOI":"10.1007\/978-3-030-58526-6_15"},{"key":"1739_CR36","doi-asserted-by":"publisher","unstructured":"Lin, Y.-J., Wu, P.-W., Chang, C.-H., Chang, E., Liao, S.-W.: Relgan: Multi-domain image-to-image translation via relative attributes. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 5913\u20135921 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00601","DOI":"10.1109\/ICCV.2019.00601"},{"key":"1739_CR37","doi-asserted-by":"publisher","unstructured":"Tewari, A., Elgharib, M., Bharaj, G., Bernard, F., Seidel, H., Perez, P., Zollhofer, M., Theobalt, C.: Stylerig: Rigging stylegan for 3d control over portrait images. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6141\u20136150 (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.00618","DOI":"10.1109\/CVPR42600.2020.00618"},{"key":"1739_CR38","doi-asserted-by":"publisher","first-page":"2004","DOI":"10.1109\/TPAMI.2020.3034267","volume":"44","author":"Y Shen","year":"2020","unstructured":"Shen, Y., Yang, C., Tang, X., Zhou, B.: Interfacegan: interpreting the disentangled face representation learned by gans. IEEE Trans. Pattern Anal. Mach. Intell. 44, 2004\u20132018 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1739_CR39","unstructured":"Mao, W., Lou, B., Yuan, J.: Tunagan: Interpretable gan for smart editing. ArXiv arXiv:abs\/1908.06163 (2019)"},{"key":"1739_CR40","doi-asserted-by":"publisher","unstructured":"Zhuang, P., Koyejo, O., Schwing, A.G.: Enjoy your editing: Controllable gans for image editing via latent space navigation. ICLR, (2021). https:\/\/doi.org\/10.48550\/arXiv.2102.01187","DOI":"10.48550\/arXiv.2102.01187"},{"key":"1739_CR41","doi-asserted-by":"crossref","unstructured":"Chan, E.R., Lin, C.Z., Chan, M.A., Nagano, K., Pan, B., Mello, S.D., Gallo, O., Guibas, L., Tremblay, J., Khamis, S., Karras, T., Wetzstein, G.: Efficient Geometry-aware 3D Generative Adversarial Networks (2022). arXiv:org\/abs\/2112.07945","DOI":"10.1109\/CVPR52688.2022.01565"},{"key":"1739_CR42","doi-asserted-by":"crossref","unstructured":"Or-El, R., Luo, X., Shan, M., Shechtman, E., Park, J.J., Kemelmacher-Shlizerman, I.: StyleSDF: High-Resolution 3D-Consistent Image and Geometry Generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13503\u201313513 (2022)","DOI":"10.1109\/CVPR52688.2022.01314"},{"key":"1739_CR43","doi-asserted-by":"publisher","unstructured":"Liu, Z., Wang, Y., Vaidya, S., Ruehle, F., Halverson, J., Solja\u010di\u0107, M., Hou, T.Y., Tegmark, M.: KAN: Kolmogorov-Arnold Networks (2024). https:\/\/doi.org\/10.48550\/arXiv.2404.19756","DOI":"10.48550\/arXiv.2404.19756"},{"issue":"86","key":"1739_CR44","first-page":"2579","volume":"9","author":"L Maaten","year":"2008","unstructured":"Maaten, L., Hinton, G.: Visualizing data using t-SNE. JMLR 9(86), 2579\u20132605 (2008)","journal-title":"JMLR"}],"container-title":["Machine Vision and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-025-01739-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00138-025-01739-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-025-01739-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,10]],"date-time":"2025-11-10T17:01:01Z","timestamp":1762794061000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00138-025-01739-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,8]]},"references-count":44,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,11]]}},"alternative-id":["1739"],"URL":"https:\/\/doi.org\/10.1007\/s00138-025-01739-z","relation":{},"ISSN":["0932-8092","1432-1769"],"issn-type":[{"type":"print","value":"0932-8092"},{"type":"electronic","value":"1432-1769"}],"subject":[],"published":{"date-parts":[[2025,9,8]]},"assertion":[{"value":"21 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 July 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 August 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 September 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"117"}}