{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T16:51:36Z","timestamp":1777654296054,"version":"3.51.4"},"publisher-location":"Cham","reference-count":63,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031726835","type":"print"},{"value":"9783031726842","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72684-2_4","type":"book-chapter","created":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T19:02:45Z","timestamp":1730574165000},"page":"52-69","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["High-Fidelity 3D Textured Shapes Generation by\u00a0Sparse Encoding and\u00a0Adversarial Decoding"],"prefix":"10.1007","author":[{"given":"Qi","family":"Zuo","sequence":"first","affiliation":[]},{"given":"Xiaodong","family":"Gu","sequence":"additional","affiliation":[]},{"given":"Yuan","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Zhengyi","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Weihao","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Lingteng","family":"Qiu","sequence":"additional","affiliation":[]},{"given":"Liefeng","family":"Bo","sequence":"additional","affiliation":[]},{"given":"Zilong","family":"Dong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,3]]},"reference":[{"key":"4_CR1","doi-asserted-by":"crossref","unstructured":"Cao, T., Kreis, K., Fidler, S., Sharp, N., Yin, K.: TexFusion: synthesizing 3D textures with text-guided image diffusion models. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4169\u20134181 (2023)","DOI":"10.1109\/ICCV51070.2023.00385"},{"key":"4_CR2","doi-asserted-by":"crossref","unstructured":"Chan, E.R., et\u00a0al.: Efficient geometry-aware 3D generative adversarial networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16123\u201316133 (2022)","DOI":"10.1109\/CVPR52688.2022.01565"},{"key":"4_CR3","doi-asserted-by":"crossref","unstructured":"Chan, E.R., Monteiro, M., Kellnhofer, P., Wu, J., Wetzstein, G.: pi-GAN: periodic implicit generative adversarial networks for 3D-aware image synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5799\u20135809 (2021)","DOI":"10.1109\/CVPR46437.2021.00574"},{"key":"4_CR4","unstructured":"Chang, A.X., et al.: ShapeNet: an information-rich 3D model repository. Technical report. arXiv:1512.03012 [cs.GR], Stanford University\u2014Princeton University\u2014Toyota Technological Institute at Chicago (2015)"},{"key":"4_CR5","doi-asserted-by":"crossref","unstructured":"Chen, D.Z., Siddiqui, Y., Lee, H.Y., Tulyakov, S., Nie\u00dfner, M.: Text2Tex: text-driven texture synthesis via diffusion models. arXiv preprint arXiv:2303.11396 (2023)","DOI":"10.1109\/ICCV51070.2023.01701"},{"key":"4_CR6","doi-asserted-by":"crossref","unstructured":"Chen, R., Chen, Y., Jiao, N., Jia, K.: Fantasia3D: disentangling geometry and appearance for high-quality text-to-3D content creation. arXiv preprint arXiv:2303.13873 (2023)","DOI":"10.1109\/ICCV51070.2023.02033"},{"key":"4_CR7","doi-asserted-by":"crossref","unstructured":"Cheng, Y.C., Lee, H.Y., Tulyakov, S., Schwing, A.G., Gui, L.Y.: SDFusion: multimodal 3D shape completion, reconstruction, and generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4456\u20134465 (2023)","DOI":"10.1109\/CVPR52729.2023.00433"},{"key":"4_CR8","doi-asserted-by":"crossref","unstructured":"Chou, G., Bahat, Y., Heide, F.: Diffusion-SDF: conditional generative modeling of signed distance functions. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2262\u20132272 (2023)","DOI":"10.1109\/ICCV51070.2023.00215"},{"key":"4_CR9","unstructured":"Chou, G., Chugunov, I., Heide, F.: GenSDF: two-stage learning of generalizable signed distance functions. In: Advances in Neural Information Processing Systems, vol. 35, pp. 24905\u201324919 (2022)"},{"key":"4_CR10","doi-asserted-by":"crossref","unstructured":"Choy, C., Gwak, J., Savarese, S.: 4D spatio-temporal convnets: minkowski convolutional neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3075\u20133084 (2019)","DOI":"10.1109\/CVPR.2019.00319"},{"key":"4_CR11","doi-asserted-by":"crossref","unstructured":"Deitke, M., et al.: Objaverse: a universe of annotated 3D objects. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13142\u201313153 (2023)","DOI":"10.1109\/CVPR52729.2023.01263"},{"key":"4_CR12","doi-asserted-by":"crossref","unstructured":"Dong, Y., et al.: GPLD3D: latent diffusion of 3d shape generative models by enforcing geometric and physical priors. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 56\u201366 (2024)","DOI":"10.1109\/CVPR52733.2024.00014"},{"key":"4_CR13","unstructured":"Gao, J., et al.: GET3D: a generative model of high quality 3D textured shapes learned from images. In: Advances in Neural Information Processing Systems, vol. 35, pp. 31841\u201331854 (2022)"},{"key":"4_CR14","unstructured":"Gupta, A., Xiong, W., Nie, Y., Jones, I., O\u011fuz, B.: 3DGen: triplane latent diffusion for textured mesh generation. arXiv preprint arXiv:2303.05371 (2023)"},{"key":"4_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"297","DOI":"10.1007\/978-3-030-58548-8_18","volume-title":"Computer Vision \u2013 ECCV 2020","author":"JY Gwak","year":"2020","unstructured":"Gwak, J.Y., Choy, C., Savarese, S.: Generative sparse detection networks for 3D single-shot object detection. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12349, pp. 297\u2013313. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58548-8_18"},{"key":"4_CR16","doi-asserted-by":"publisher","unstructured":"Jakob, W., Speierer, S., Roussel, N., Vicini, D.: Dr.jit: a just-in-time compiler for differentiable rendering. Trans. Graph. (Proceedings of SIGGRAPH) 41(4) (2022). https:\/\/doi.org\/10.1145\/3528223.3530099","DOI":"10.1145\/3528223.3530099"},{"key":"4_CR17","unstructured":"Jun, H., Nichol, A.: Shap-E: generating conditional 3D implicit functions. arXiv preprint arXiv:2305.02463 (2023)"},{"key":"4_CR18","doi-asserted-by":"crossref","unstructured":"Kang, M., et al.: Scaling up GANs for text-to-image synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10124\u201310134 (2023)","DOI":"10.1109\/CVPR52729.2023.00976"},{"key":"4_CR19","doi-asserted-by":"crossref","unstructured":"Laine, S., Hellsten, J., Karras, T., Seol, Y., Lehtinen, J., Aila, T.: Modular primitives for high-performance differentiable rendering. ACM Trans. Graph. 39(6) (2020)","DOI":"10.1145\/3414685.3417861"},{"key":"4_CR20","unstructured":"Li, W., Chen, R., Chen, X., Tan, P.: SweetDreamer: aligning geometric priors in 2D diffusion for consistent text-to-3D. arXiv preprint arXiv:2310.02596 (2023)"},{"key":"4_CR21","doi-asserted-by":"crossref","unstructured":"Lin, C.H., et al.: Magic3D: high-resolution text-to-3D content creation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 300\u2013309 (2023)","DOI":"10.1109\/CVPR52729.2023.00037"},{"key":"4_CR22","unstructured":"Liu, M., et al.: OpenShape: scaling up 3D shape representation towards open-world understanding. arXiv preprint arXiv:2305.10764 (2023)"},{"key":"4_CR23","doi-asserted-by":"crossref","unstructured":"Liu, R., Wu, R., Van\u00a0Hoorick, B., Tokmakov, P., Zakharov, S., Vondrick, C.: Zero-1-to-3: zero-shot one image to 3D object. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9298\u20139309 (2023)","DOI":"10.1109\/ICCV51070.2023.00853"},{"key":"4_CR24","unstructured":"Liu, Y., et al.: SyncDreamer: generating multiview-consistent images from a single-view image. arXiv preprint arXiv:2309.03453 (2023)"},{"key":"4_CR25","unstructured":"Liu, Z., Tang, H., Lin, Y., Han, S.: Point-voxel CNN for efficient 3D deep learning. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"4_CR26","doi-asserted-by":"crossref","unstructured":"Long, X., et\u00a0al.: Wonder3D: single image to 3d using cross-domain diffusion. arXiv preprint arXiv:2310.15008 (2023)","DOI":"10.1109\/CVPR52733.2024.00951"},{"key":"4_CR27","doi-asserted-by":"crossref","unstructured":"Mao, J., et al.: Voxel transformer for 3D object detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3164\u20133173 (2021)","DOI":"10.1109\/ICCV48922.2021.00315"},{"key":"4_CR28","doi-asserted-by":"crossref","unstructured":"Mescheder, L., Oechsle, M., Niemeyer, M., Nowozin, S., Geiger, A.: Occupancy networks: learning 3D reconstruction in function space. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4460\u20134470 (2019)","DOI":"10.1109\/CVPR.2019.00459"},{"key":"4_CR29","doi-asserted-by":"crossref","unstructured":"Mittal, P., Cheng, Y.C., Singh, M., Tulsiani, S.: AutoSDF: shape priors for 3D completion, reconstruction and generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 306\u2013315 (2022)","DOI":"10.1109\/CVPR52688.2022.00040"},{"key":"4_CR30","doi-asserted-by":"crossref","unstructured":"Nguyen-Phuoc, T., Li, C., Theis, L., Richardt, C., Yang, Y.L.: HoloGAN: unsupervised learning of 3D representations from natural images. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7588\u20137597 (2019)","DOI":"10.1109\/ICCV.2019.00768"},{"key":"4_CR31","unstructured":"Nichol, A., Jun, H., Dhariwal, P., Mishkin, P., Chen, M.: Point-e: a system for generating 3D point clouds from complex prompts. arXiv preprint arXiv:2212.08751 (2022)"},{"key":"4_CR32","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Geiger, A.: GIRAFFE: representing scenes as compositional generative neural feature fields. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11453\u201311464 (2021)","DOI":"10.1109\/CVPR46437.2021.01129"},{"key":"4_CR33","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"523","DOI":"10.1007\/978-3-030-58580-8_31","volume-title":"Computer Vision \u2013 ECCV 2020","author":"S Peng","year":"2020","unstructured":"Peng, S., Niemeyer, M., Mescheder, L., Pollefeys, M., Geiger, A.: Convolutional occupancy networks. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12348, pp. 523\u2013540. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58580-8_31"},{"key":"4_CR34","unstructured":"Poole, B., Jain, A., Barron, J.T., Mildenhall, B.: DreamFusion: text-to-3D using 2D diffusion. arXiv preprint arXiv:2209.14988 (2022)"},{"key":"4_CR35","doi-asserted-by":"crossref","unstructured":"Qiu, L., et al.: RichDreamer: a generalizable normal-depth diffusion model for detail richness in text-to-3D (2023). https:\/\/arxiv.org\/abs\/2311.16918","DOI":"10.1109\/CVPR52733.2024.00946"},{"key":"4_CR36","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"4_CR37","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"4_CR38","unstructured":"Saharia, C., et al.: Photorealistic text-to-image diffusion models with deep language understanding. In: Advances in Neural Information Processing Systems, vol. 35, pp. 36479\u201336494 (2022)"},{"key":"4_CR39","doi-asserted-by":"crossref","unstructured":"Sander, P.V., Snyder, J., Gortler, S.J., Hoppe, H.: Texture mapping progressive meshes. In: Proceedings of the 28th Annual Conference on Computer Graphics and Interactive Techniques, pp. 409\u2013416 (2001)","DOI":"10.1145\/383259.383307"},{"key":"4_CR40","doi-asserted-by":"crossref","unstructured":"Sanghi, A., et al.: Clip-forge: towards zero-shot text-to-shape generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18603\u201318613 (2022)","DOI":"10.1109\/CVPR52688.2022.01805"},{"issue":"4","key":"4_CR41","first-page":"1","volume":"42","author":"T Shen","year":"2023","unstructured":"Shen, T., et al.: Flexible isosurface extraction for gradient-based mesh optimization. ACM Trans. Graph. (TOG) 42(4), 1\u201316 (2023)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"4_CR42","unstructured":"Shi, R., et al.: Zero123++: a single image to consistent multi-view diffusion base model. arXiv preprint arXiv:2310.15110 (2023)"},{"key":"4_CR43","unstructured":"Shi, Y., Wang, P., Ye, J., Long, M., Li, K., Yang, X.: MVDream: multi-view diffusion for 3D generation. arXiv preprint arXiv:2308.16512 (2023)"},{"key":"4_CR44","doi-asserted-by":"crossref","unstructured":"Shue, J.R., Chan, E.R., Po, R., Ankner, Z., Wu, J., Wetzstein, G.: 3D neural field generation using triplane diffusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20875\u201320886 (2023)","DOI":"10.1109\/CVPR52729.2023.02000"},{"key":"4_CR45","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1007\/978-3-031-20062-5_5","volume-title":"ECCV 2022","author":"Y Siddiqui","year":"2022","unstructured":"Siddiqui, Y., Thies, J., Ma, F., Shan, Q., Nie\u00dfner, M., Dai, A.: Texturify: generating textures on 3D shape surfaces. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13663, pp. 72\u201388. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20062-5_5"},{"key":"4_CR46","unstructured":"Song, J., Meng, C., Ermon, S.: Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502 (2020)"},{"key":"4_CR47","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"685","DOI":"10.1007\/978-3-030-58604-1_41","volume-title":"Computer Vision \u2013 ECCV 2020","author":"H Tang","year":"2020","unstructured":"Tang, H., et al.: Searching efficient 3D architectures with sparse point-voxel convolution. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12373, pp. 685\u2013702. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58604-1_41"},{"key":"4_CR48","unstructured":"Tang, J., Ren, J., Zhou, H., Liu, Z., Zeng, G.: DreamGaussian: generative gaussian splatting for efficient 3D content creation. arXiv preprint arXiv:2309.16653 (2023)"},{"key":"4_CR49","unstructured":"Vahdat, A., Williams, F., Gojcic, Z., Litany, O., Fidler, S., Kreis, K., et al.: Lion: latent point diffusion models for 3D shape generation. In: Advances in Neural Information Processing Systems, vol. 35, pp. 10021\u201310039 (2022)"},{"issue":"4","key":"4_CR50","first-page":"1","volume":"36","author":"PS Wang","year":"2017","unstructured":"Wang, P.S., Liu, Y., Guo, Y.X., Sun, C.Y., Tong, X.: O-CNN: octree-based convolutional neural networks for 3D shape analysis. ACM Trans. Graph. (TOG) 36(4), 1\u201311 (2017)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"4_CR51","unstructured":"Wang, Z., et al.: ProlificDreamer: high-fidelity and diverse text-to-3D generation with variational score distillation. arXiv preprint arXiv:2305.16213 (2023)"},{"key":"4_CR52","unstructured":"Weng, H., et al.: Consistent123: improve consistency for one image to 3D object synthesis. arXiv preprint arXiv:2310.08092 (2023)"},{"key":"4_CR53","doi-asserted-by":"crossref","unstructured":"Yang, G., Huang, X., Hao, Z., Liu, M.Y., Belongie, S., Hariharan, B.: PointFlow: 3D point cloud generation with continuous normalizing flows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4541\u20134550 (2019)","DOI":"10.1109\/ICCV.2019.00464"},{"key":"4_CR54","unstructured":"Zeng, X., et al.: Lion: latent point diffusion models for 3D shape generation. arXiv preprint arXiv:2210.06978 (2022)"},{"key":"4_CR55","unstructured":"Zhang, B., Nie\u00dfner, M., Wonka, P.: 3DILG: irregular latent grids for 3D generative modeling. In: Advances in Neural Information Processing Systems, vol. 35, pp. 21871\u201321885 (2022)"},{"key":"4_CR56","doi-asserted-by":"crossref","unstructured":"Zhang, B., Tang, J., Niessner, M., Wonka, P.: 3DShape2VecSet: a 3D shape representation for neural fields and generative diffusion models (2023). https:\/\/arxiv.org\/abs\/2301.11445","DOI":"10.1145\/3592442"},{"key":"4_CR57","doi-asserted-by":"crossref","unstructured":"Zhang, P., Zhang, B., Chen, D., Yuan, L., Wen, F.: Cross-domain correspondence learning for exemplar-based image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5143\u20135153 (2020)","DOI":"10.1109\/CVPR42600.2020.00519"},{"key":"4_CR58","doi-asserted-by":"crossref","unstructured":"Zhao, Z., et al.: An optimization framework to enforce multi-view consistency for texturing 3D meshes using pre-trained text-to-image models (2024). https:\/\/arxiv.org\/abs\/2403.15559","DOI":"10.1007\/978-3-031-72764-1_9"},{"key":"4_CR59","doi-asserted-by":"crossref","unstructured":"Zheng, X.Y., Pan, H., Wang, P.S., Tong, X., Liu, Y., Shum, H.Y.: Locally attentional SDF diffusion for controllable 3D shape generation. arXiv preprint arXiv:2305.04461 (2023)","DOI":"10.1145\/3592103"},{"key":"4_CR60","doi-asserted-by":"crossref","unstructured":"Zhou, L., Du, Y., Wu, J.: 3D shape generation and completion through point-voxel diffusion. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5826\u20135835 (2021)","DOI":"10.1109\/ICCV48922.2021.00577"},{"key":"4_CR61","doi-asserted-by":"crossref","unstructured":"Zhou, X., et al.: CoCosNet v2: full-resolution correspondence learning for image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11465\u201311475 (2021)","DOI":"10.1109\/CVPR46437.2021.01130"},{"key":"4_CR62","unstructured":"Zuo, Q., et al.: VideoMV: consistent multi-view generation based on large video generative model (2024). https:\/\/arxiv.org\/abs\/2403.12010"},{"key":"4_CR63","doi-asserted-by":"publisher","unstructured":"Zuo, Q., Song, Y., Li, J., Liu, L., Bo, L.: DG3D: generating high quality 3D textured shapes by learning to discriminate multi-modal diffusion-renderings. In: 2023 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 14529\u201314538 (2023). https:\/\/doi.org\/10.1109\/ICCV51070.2023.01340","DOI":"10.1109\/ICCV51070.2023.01340"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72684-2_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T19:03:04Z","timestamp":1730574184000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72684-2_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,3]]},"ISBN":["9783031726835","9783031726842"],"references-count":63,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72684-2_4","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,3]]},"assertion":[{"value":"3 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}