{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T15:39:18Z","timestamp":1775230758094,"version":"3.50.1"},"publisher-location":"Cham","reference-count":55,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031726484","type":"print"},{"value":"9783031726491","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T00:00:00Z","timestamp":1727654400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T00:00:00Z","timestamp":1727654400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72649-1_3","type":"book-chapter","created":{"date-parts":[[2024,9,29]],"date-time":"2024-09-29T07:01:50Z","timestamp":1727593310000},"page":"37-53","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":38,"title":["MVSGaussian: Fast Generalizable Gaussian Splatting Reconstruction from\u00a0Multi-View Stereo"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-0718-0614","authenticated-orcid":false,"given":"Tianqi","family":"Liu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6627-814X","authenticated-orcid":false,"given":"Guangcong","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3345-6923","authenticated-orcid":false,"given":"Shoukang","family":"Hu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2423-4835","authenticated-orcid":false,"given":"Liao","family":"Shen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0009-1126-3336","authenticated-orcid":false,"given":"Xinyi","family":"Ye","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1110-5062","authenticated-orcid":false,"given":"Yuhang","family":"Zang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9223-1863","authenticated-orcid":false,"given":"Zhiguo","family":"Cao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0008-0600-3794","authenticated-orcid":false,"given":"Wei","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4220-5958","authenticated-orcid":false,"given":"Ziwei","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,30]]},"reference":[{"key":"3_CR1","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1007\/s11263-016-0902-9","volume":"120","author":"H Aanaes","year":"2016","unstructured":"Aanaes, H., Jensen, R.R., Vogiatzis, G., Tola, E., Dahl, A.B.: Large-scale data for multiple-view stereopsis. IJCV 120, 153\u2013168 (2016)","journal-title":"IJCV"},{"key":"3_CR2","doi-asserted-by":"crossref","unstructured":"Boss, M., Braun, R., Jampani, V., Barron, J.T., Liu, C., Lensch, H.: NeRD: neural reflectance decomposition from image collections. In: ICCV, pp. 12684\u201312694 (2021)","DOI":"10.1109\/ICCV48922.2021.01245"},{"key":"3_CR3","unstructured":"Cen, J., et al.: Segment any 3D Gaussians. arXiv preprint arXiv:2312.00860 (2023)"},{"key":"3_CR4","doi-asserted-by":"crossref","unstructured":"Charatan, D., Li, S., Tagliasacchi, A., Sitzmann, V.: pixelSplat: 3D Gaussian splats from image pairs for scalable generalizable 3D reconstruction. In: arXiv (2023)","DOI":"10.1109\/CVPR52733.2024.01840"},{"issue":"1","key":"3_CR5","first-page":"1","volume":"41","author":"A Chen","year":"2022","unstructured":"Chen, A., Liu, R., Xie, L., Chen, Z., Su, H., Yu, J.: SofGAN: a portrait image generator with dynamic styling. ACM Trans. Graph. 41(1), 1\u201326 (2022)","journal-title":"ACM Trans. Graph."},{"key":"3_CR6","doi-asserted-by":"crossref","unstructured":"Chen, A., et al.: MVSNeRF: fast generalizable radiance field reconstruction from multi-view stereo. In: ICCV, pp. 14124\u201314133 (2021)","DOI":"10.1109\/ICCV48922.2021.01386"},{"key":"3_CR7","unstructured":"Chen, G., Wang, W.: A survey on 3D Gaussian splatting. arXiv preprint arXiv:2401.03890 (2024)"},{"key":"3_CR8","doi-asserted-by":"crossref","unstructured":"Chen, Y., et al.: GaussianEditor: swift and controllable 3D editing with gaussian splatting. arXiv preprint arXiv:2311.14521 (2023)","DOI":"10.1109\/CVPR52733.2024.02029"},{"key":"3_CR9","unstructured":"Chen, Y., Xu, H., Wu, Q., Zheng, C., Cham, T.J., Cai, J.: Explicit correspondence matching for generalizable neural radiance fields. arXiv preprint arXiv:2304.12294 (2023)"},{"key":"3_CR10","doi-asserted-by":"crossref","unstructured":"Cheng, S., et al.: Deep stereo using adaptive thin volume representation with uncertainty awareness. In: CVPR, pp. 2524\u20132534 (2020)","DOI":"10.1109\/CVPR42600.2020.00260"},{"key":"3_CR11","doi-asserted-by":"crossref","unstructured":"Ding, Y., et al.: TransMVSNet global context-aware multi-view stereo network with transformers. In: CVPR, pp. 8585\u20138594 (2022)","DOI":"10.1109\/CVPR52688.2022.00839"},{"key":"3_CR12","doi-asserted-by":"crossref","unstructured":"Fua, P., Leclerc, Y.G.: Object-centered surface reconstruction combining multi-image stereo and shading. IJCV 16, 35\u201356 (1995)","DOI":"10.1007\/BF01428192"},{"key":"3_CR13","doi-asserted-by":"crossref","unstructured":"Galliani, S., Lasinger, K., Schindler, K.: Massively parallel multiview stereopsis by surface normal diffusion. In: ICCV, pp. 873\u2013881 (2015)","DOI":"10.1109\/ICCV.2015.106"},{"key":"3_CR14","doi-asserted-by":"crossref","unstructured":"Gu, X., Fan, Z., Zhu, S., Dai, Z., Tan, F., Tan, P.: Cascade cost volume for high-resolution multi-view stereo and stereo matching. In: CVPR, pp. 2495\u20132504 (2020)","DOI":"10.1109\/CVPR42600.2020.00257"},{"key":"3_CR15","doi-asserted-by":"crossref","unstructured":"Hu, L., et al.: GaussianAvatar: towards realistic human avatar modeling from a single video via animatable 3D Gaussians. arXiv preprint arXiv:2312.02134 (2023)","DOI":"10.1109\/CVPR52733.2024.00067"},{"key":"3_CR16","doi-asserted-by":"crossref","unstructured":"Hu, S., Liu, Z.: GauHuman: articulated gaussian splatting from monocular human videos. arXiv preprint arXiv: (2023)","DOI":"10.1109\/CVPR52733.2024.01930"},{"key":"3_CR17","unstructured":"Hu, S., et al.: ConsistentNeRF: enhancing neural radiance fields with 3D consistency for sparse view synthesis. arXiv preprint arXiv:2305.11031 (2023)"},{"key":"3_CR18","doi-asserted-by":"crossref","unstructured":"Irshad, M.Z., et al.: Neo 360: neural fields for sparse view synthesis of outdoor scenes (2023)","DOI":"10.1109\/ICCV51070.2023.00843"},{"issue":"4","key":"3_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3592433","volume":"42","author":"B Kerbl","year":"2023","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., Drettakis, G.: 3D Gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 42(4), 1\u201314 (2023)","journal-title":"ACM Trans. Graph."},{"key":"3_CR20","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"issue":"4","key":"3_CR21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073599","volume":"36","author":"A Knapitsch","year":"2017","unstructured":"Knapitsch, A., Park, J., Zhou, Q.Y., Koltun, V.: Tanks and temples benchmarking large-scale scene reconstruction. ACM Trans. Graph. 36(4), 1\u201313 (2017)","journal-title":"ACM Trans. Graph."},{"key":"3_CR22","doi-asserted-by":"crossref","unstructured":"Lin, H., et al.: Efficient neural radiance fields for interactive free-viewpoint video. In: SIGGRAPH Asia Conference Proceedings (2022)","DOI":"10.1145\/3550469.3555376"},{"key":"3_CR23","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: CVPR, pp. 2117\u20132125 (2017)","DOI":"10.1109\/CVPR.2017.106"},{"key":"3_CR24","doi-asserted-by":"crossref","unstructured":"Liu, T., et al.: Geometry-aware reconstruction and fusion-refined rendering for generalizable neural radiance fields. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7654\u20137663, June 2024","DOI":"10.1109\/CVPR52733.2024.00731"},{"key":"3_CR25","doi-asserted-by":"crossref","unstructured":"Liu, T., Ye, X., Zhao, W., Pan, Z., Shi, M., Cao, Z.: When epipolar constraint meets non-local operators in multi-view stereo. In: ICCV, pp. 18088\u201318097 (2023)","DOI":"10.1109\/ICCV51070.2023.01658"},{"key":"3_CR26","doi-asserted-by":"crossref","unstructured":"Liu, Y., et al.: Neural rays for occlusion-aware image-based rendering. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00767"},{"key":"3_CR27","doi-asserted-by":"crossref","unstructured":"Luiten, J., Kopanas, G., Leibe, B., Ramanan, D.: Dynamic 3D Gaussians: tracking by persistent dynamic view synthesis. In: 3DV (2024)","DOI":"10.1109\/3DV62453.2024.00044"},{"issue":"4","key":"3_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3306346.3322980","volume":"38","author":"B Mildenhall","year":"2019","unstructured":"Mildenhall, B., et al.: Local light field fusion: practical view synthesis with prescriptive sampling guidelines. ACM Trans. Graph. 38(4), 1\u201314 (2019)","journal-title":"ACM Trans. Graph."},{"key":"3_CR29","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"3_CR30","doi-asserted-by":"crossref","unstructured":"Park, K., et al.: Nerfies: Deformable neural radiance fields. In: ICCV, pp. 5865\u20135874 (2021)","DOI":"10.1109\/ICCV48922.2021.00581"},{"key":"3_CR31","doi-asserted-by":"crossref","unstructured":"Peng, R., Wang, R., Wang, Z., Lai, Y., Wang, R.: Rethinking depth estimation for multi-view stereo a unified representation. In: CVPR, pp. 8645\u20138654 (2022)","DOI":"10.1109\/CVPR52688.2022.00845"},{"key":"3_CR32","doi-asserted-by":"crossref","unstructured":"Qian, Z., Wang, S., Mihajlovic, M., Geiger, A., Tang, S.: 3DGS-Avatar: animatable avatars via deformable 3D Gaussian splatting. arXiv preprint arXiv:2312.09228 (2023)","DOI":"10.1109\/CVPR52733.2024.00480"},{"key":"3_CR33","doi-asserted-by":"crossref","unstructured":"Schonberger, J.L., Frahm, J.M.: Structure-from-motion revisited. In: CVPR, pp. 4104\u20134113 (2016)","DOI":"10.1109\/CVPR.2016.445"},{"key":"3_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"501","DOI":"10.1007\/978-3-319-46487-9_31","volume-title":"Computer Vision \u2013 ECCV 2016","author":"JL Sch\u00f6nberger","year":"2016","unstructured":"Sch\u00f6nberger, J.L., Zheng, E., Frahm, J.-M., Pollefeys, M.: Pixelwise view selection for unstructured multi-view stereo. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9907, pp. 501\u2013518. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46487-9_31"},{"key":"3_CR35","doi-asserted-by":"crossref","unstructured":"Szymanowicz, S., Rupprecht, C., Vedaldi, A.: Splatter image: ultra-fast single-view 3D reconstruction. In: arXiv (2023)","DOI":"10.1109\/CVPR52733.2024.00972"},{"key":"3_CR36","unstructured":"Wang, P., Chen, X., Chen, T., Venugopalan, S., Wang, Z.: Is attention all that neRF needs? In: ICLR (2023)"},{"key":"3_CR37","unstructured":"Vaswani, A., et al.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"3_CR38","doi-asserted-by":"crossref","unstructured":"Wang, G., Chen, Z., Loy, C.C., Liu, Z.: SparseNeRF: distilling depth ranking for few-shot novel view synthesis. In: IEEE\/CVF International Conference on Computer Vision (ICCV) (2023)","DOI":"10.1109\/ICCV51070.2023.00832"},{"key":"3_CR39","doi-asserted-by":"publisher","first-page":"6905","DOI":"10.1109\/TPAMI.2024.3387307","volume":"46","author":"G Wang","year":"2024","unstructured":"Wang, G., Wang, P., Chen, Z., Wang, W., Loy, C.C., Liu, Z.: PERF: panoramic neural radiance field from a single panorama. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 46, 6905\u20136918 (2024)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI)"},{"key":"3_CR40","doi-asserted-by":"crossref","unstructured":"Wang, Q., et al.: IBRNet: learning multi-view image-based rendering. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00466"},{"key":"3_CR41","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"573","DOI":"10.1007\/978-3-031-19821-2_33","volume-title":"ECCV 2022","author":"X Wang","year":"2022","unstructured":"Wang, X., et al.: MVSTER epipolar transformer for efficient multi-view stereo. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. Lecture Notes in Computer Science, vol. 13691, pp. 573\u2013591. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19821-2_33"},{"issue":"4","key":"3_CR42","first-page":"600","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE TIP 13(4), 600\u2013612 (2004)","journal-title":"IEEE TIP"},{"key":"3_CR43","doi-asserted-by":"crossref","unstructured":"Wei, Z., Zhu, Q., Min, C., Chen, Y., Wang, G.: AA-RMVSNet adaptive aggregation recurrent multi-view stereo network. In: ICCV, pp. 6187\u20136196 (2021)","DOI":"10.1109\/ICCV48922.2021.00613"},{"key":"3_CR44","doi-asserted-by":"crossref","unstructured":"Wu, G., et al.: 4D Gaussian splatting for real-time dynamic scene rendering. arXiv preprint arXiv:2310.08528 (2023)","DOI":"10.1109\/CVPR52733.2024.01920"},{"key":"3_CR45","doi-asserted-by":"crossref","unstructured":"Xian, W., Huang, J.B., Kopf, J., Kim, C.: Space-time neural irradiance fields for free-viewpoint video. In: CVPR, pp. 9421\u20139431 (2021)","DOI":"10.1109\/CVPR46437.2021.00930"},{"key":"3_CR46","doi-asserted-by":"crossref","unstructured":"Xiang, F., Xu, Z., Hasan, M., Hold-Geoffroy, Y., Sunkavalli, K., Su, H.: NeuTex: neural texture mapping for volumetric neural rendering. In: CVPR, pp. 7119\u20137128 (2021)","DOI":"10.1109\/CVPR46437.2021.00704"},{"key":"3_CR47","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"674","DOI":"10.1007\/978-3-030-58548-8_39","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Yan","year":"2020","unstructured":"Yan, J., et al.: Dense hybrid recurrent multi-view stereo net with dynamic consistency checking. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12349, pp. 674\u2013689. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58548-8_39"},{"key":"3_CR48","doi-asserted-by":"crossref","unstructured":"Yang, J., Mao, W., Alvarez, J.M., Liu, M.: Cost volume pyramid based depth inference for multi-view stereo. In: CVPR, pp. 4877\u20134886 (2020)","DOI":"10.1109\/CVPR42600.2020.00493"},{"key":"3_CR49","doi-asserted-by":"crossref","unstructured":"Yang, Z., Gao, X., Zhou, W., Jiao, S., Zhang, Y., Jin, X.: Deformable 3d gaussians for high-fidelity monocular dynamic scene reconstruction. arXiv preprint arXiv:2309.13101 (2023)","DOI":"10.1109\/CVPR52733.2024.01922"},{"key":"3_CR50","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"785","DOI":"10.1007\/978-3-030-01237-3_47","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Yao","year":"2018","unstructured":"Yao, Y., Luo, Z., Li, S., Fang, T., Quan, L.: MVSNet: depth inference for unstructured multi-view stereo. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11212, pp. 785\u2013801. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01237-3_47"},{"key":"3_CR51","doi-asserted-by":"crossref","unstructured":"Yao, Y., Luo, Z., Li, S., Shen, T., Fang, T., Quan, L.: Recurrent MVSNet for high-resolution multi-view stereo depth inference. In: CVPR, pp. 5525\u20135534 (2019)","DOI":"10.1109\/CVPR.2019.00567"},{"key":"3_CR52","doi-asserted-by":"crossref","unstructured":"Ye, X., Zhao, W., Liu, T., Huang, Z., Cao, Z., Li, X.: Constraining depth map geometry for multi-view stereo: A dual-depth approach with saddle-shaped depth cells. In: ICCV, pp. 17661\u201317670 (2023)","DOI":"10.1109\/ICCV51070.2023.01619"},{"key":"3_CR53","doi-asserted-by":"crossref","unstructured":"Yu, A., Ye, V., Tancik, M., Kanazawa, A.: pixelNeRF: neural radiance fields from one or few images. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"3_CR54","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"3_CR55","doi-asserted-by":"crossref","unstructured":"Zheng, S., et al.: GPS-Gaussian: generalizable pixel-wise 3d gaussian splatting for real-time human novel view synthesis. arXiv (2023)","DOI":"10.1109\/CVPR52733.2024.01861"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72649-1_3","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T21:18:22Z","timestamp":1732828702000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72649-1_3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,30]]},"ISBN":["9783031726484","9783031726491"],"references-count":55,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72649-1_3","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,9,30]]},"assertion":[{"value":"30 September 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}