{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,24]],"date-time":"2026-02-24T18:17:27Z","timestamp":1771957047989,"version":"3.50.1"},"publisher-location":"Cham","reference-count":62,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031732287","type":"print"},{"value":"9783031732294","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,10,25]],"date-time":"2024-10-25T00:00:00Z","timestamp":1729814400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,25]],"date-time":"2024-10-25T00:00:00Z","timestamp":1729814400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73229-4_22","type":"book-chapter","created":{"date-parts":[[2024,10,24]],"date-time":"2024-10-24T15:03:09Z","timestamp":1729782189000},"page":"378-395","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Explorative Inbetweening of\u00a0Time and\u00a0Space"],"prefix":"10.1007","author":[{"given":"Haiwen","family":"Feng","sequence":"first","affiliation":[]},{"given":"Zheng","family":"Ding","sequence":"additional","affiliation":[]},{"given":"Zhihao","family":"Xia","sequence":"additional","affiliation":[]},{"given":"Simon","family":"Niklaus","sequence":"additional","affiliation":[]},{"given":"Victoria","family":"Abrevaya","sequence":"additional","affiliation":[]},{"given":"Michael J.","family":"Black","sequence":"additional","affiliation":[]},{"given":"Xuaner","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,25]]},"reference":[{"key":"22_CR1","unstructured":"Pexels. https:\/\/www.pexels.com\/. Accessed 01 Feb 2024"},{"key":"22_CR2","unstructured":"Bar-Tal, O., et al.: Lumiere: a space-time diffusion model for video generation. arXiv preprint arXiv:2401.12945 (2024)"},{"key":"22_CR3","unstructured":"Bar-Tal, O., Yariv, L., Lipman, Y., Dekel, T.: Multidiffusion: fusing diffusion paths for controlled image generation (2023)"},{"key":"22_CR4","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: MIP-nerf 360: unbounded anti-aliased neural radiance fields. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5470\u20135479 (2022)","DOI":"10.1109\/CVPR52688.2022.00539"},{"key":"22_CR5","unstructured":"Blattmann, A., et al.: Stable video diffusion: scaling latent video diffusion models to large datasets. arXiv preprint arXiv:2311.15127 (2023)"},{"key":"22_CR6","doi-asserted-by":"crossref","unstructured":"Blattmann, A., et al.: Align your latents: high-resolution video synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 22563\u201322575 (2023)","DOI":"10.1109\/CVPR52729.2023.02161"},{"key":"22_CR7","doi-asserted-by":"crossref","unstructured":"Cai, S., et al.: Diffdreamer: towards consistent unsupervised single-view scene extrapolation with conditional diffusion models. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00204"},{"key":"22_CR8","unstructured":"Chen, H., et al.: Videocrafter1: open diffusion models for high-quality video generation (2023)"},{"issue":"3","key":"22_CR9","doi-asserted-by":"publisher","first-page":"603","DOI":"10.1109\/30.883418","volume":"46","author":"BT Choi","year":"2000","unstructured":"Choi, B.T., Lee, S.H., Ko, S.J.: New frame rate up-conversion using bi-directional motion estimation. IEEE Trans. Consum. Electron. 46(3), 603\u2013609 (2000)","journal-title":"IEEE Trans. Consum. Electron."},{"key":"22_CR10","doi-asserted-by":"crossref","unstructured":"Choi, J., Kim, S., Jeong, Y., Gwon, Y., Yoon, S.: ILVR: conditioning method for denoising diffusion probabilistic models. arXiv preprint arXiv:2108.02938 (2021)","DOI":"10.1109\/ICCV48922.2021.01410"},{"key":"22_CR11","doi-asserted-by":"crossref","unstructured":"Debevec, P.E., Taylor, C.J., Malik, J., Levin, G., Borshukov, G., Yu, Y.: Image-based modeling and rendering of architecture with interactive photogrammetry and view-dependent texture mapping. In: 1998 IEEE International Symposium on Circuits and Systems (ISCAS), vol.\u00a05, pp. 514\u2013517. IEEE (1998)","DOI":"10.1109\/ISCAS.1998.694545"},{"key":"22_CR12","doi-asserted-by":"crossref","unstructured":"Deng, K., Liu, A., Zhu, J.Y., Ramanan, D.: Depth-supervised nerf: fewer views and faster training for free. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12882\u201312891 (2022)","DOI":"10.1109\/CVPR52688.2022.01254"},{"key":"22_CR13","doi-asserted-by":"crossref","unstructured":"Du, Y., Smith, C., Tewari, A., Sitzmann, V.: Learning to render novel views from wide-baseline stereo pairs. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4970\u20134980 (2023)","DOI":"10.1109\/CVPR52729.2023.00481"},{"key":"22_CR14","doi-asserted-by":"crossref","unstructured":"Esser, P., Chiu, J., Atighehchian, P., Granskog, J., Germanidis, A.: Structure and content-guided video synthesis with diffusion models. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7346\u20137356 (2023)","DOI":"10.1109\/ICCV51070.2023.00675"},{"key":"22_CR15","doi-asserted-by":"crossref","unstructured":"Ge, S., et al.: Preserve your own correlation: a noise prior for video diffusion models. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 22930\u201322941 (2023)","DOI":"10.1109\/ICCV51070.2023.02096"},{"key":"22_CR16","unstructured":"Gu, J., et al.: Nerfdiff: single-image view synthesis with nerf-guided distillation from 3D-aware diffusion. In: International Conference on Machine Learning, pp. 11808\u201311826. PMLR (2023)"},{"key":"22_CR17","unstructured":"Guo, Y., et al.: Animatediff: animate your personalized text-to-image diffusion models without specific tuning. arXiv preprint arXiv:2307.04725 (2023)"},{"issue":"2","key":"22_CR18","doi-asserted-by":"publisher","first-page":"752","DOI":"10.1109\/TCE.2004.1309458","volume":"50","author":"T Ha","year":"2004","unstructured":"Ha, T., Lee, S., Kim, J.: Motion compensated frame interpolation by new block-based motion estimation algorithm. IEEE Trans. Consum. Electron. 50(2), 752\u2013759 (2004)","journal-title":"IEEE Trans. Consum. Electron."},{"issue":"6","key":"22_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3272127.3275084","volume":"37","author":"P Hedman","year":"2018","unstructured":"Hedman, P., Philip, J., Price, T., Frahm, J.M., Drettakis, G., Brostow, G.: Deep blending for free-viewpoint image-based rendering. ACM Trans. Graph. (ToG) 37(6), 1\u201315 (2018)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"22_CR20","unstructured":"Hertz, A., Mokady, R., Tenenbaum, J., Aberman, K., Pritch, Y., Cohen-Or, D.: Prompt-to-prompt image editing with cross attention control. arXiv preprint arXiv:2208.01626 (2022)"},{"key":"22_CR21","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"22_CR22","unstructured":"Ho, J., et al.: Imagen video: high definition video generation with diffusion models. arXiv preprint arXiv:2210.02303 (2022)"},{"key":"22_CR23","doi-asserted-by":"crossref","unstructured":"Holynski, A., Curless, B.L., Seitz, S.M., Szeliski, R.: Animating pictures with eulerian motion fields. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5810\u20135819 (2021)","DOI":"10.1109\/CVPR46437.2021.00575"},{"key":"22_CR24","unstructured":"Hu, L., Gao, X., Zhang, P., Sun, K., Zhang, B., Bo, L.: Animate anyone: consistent and controllable image-to-video synthesis for character animation. arXiv preprint arXiv:2311.17117 (2023)"},{"key":"22_CR25","unstructured":"Karras, T., Aittala, M., Aila, T., Laine, S.: Elucidating the design space of diffusion-based generative models. In: Advances in Neural Information Processing Systems, vol. 35, pp. 26565\u201326577 (2022)"},{"key":"22_CR26","doi-asserted-by":"crossref","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., Drettakis, G.: 3D gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 42(4) (2023)","DOI":"10.1145\/3592433"},{"key":"22_CR27","doi-asserted-by":"crossref","unstructured":"Khachatryan, L., et al.: Text2video-zero: text-to-image diffusion models are zero-shot video generators. arXiv preprint arXiv:2303.13439 (2023)","DOI":"10.1109\/ICCV51070.2023.01462"},{"key":"22_CR28","doi-asserted-by":"crossref","unstructured":"Li, X., Cao, Z., Sun, H., Zhang, J., Xian, K., Lin, G.: 3D cinemagraphy from a single image. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4595\u20134605 (2023)","DOI":"10.1109\/CVPR52729.2023.00446"},{"key":"22_CR29","doi-asserted-by":"crossref","unstructured":"Li, Z., Tucker, R., Snavely, N., Holynski, A.: Generative image dynamics. arXiv preprint arXiv:2309.07906 (2023)","DOI":"10.1109\/CVPR52733.2024.02279"},{"key":"22_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1007\/978-3-030-66823-5_3","volume-title":"Computer Vision \u2013 ECCV 2020 Workshops","author":"Y Liu","year":"2020","unstructured":"Liu, Y., Xie, L., Siyao, L., Sun, W., Qiao, Yu., Dong, C.: Enhanced quadratic video interpolation. In: Bartoli, A., Fusiello, A. (eds.) ECCV 2020. LNCS, vol. 12538, pp. 41\u201356. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-66823-5_3"},{"key":"22_CR31","doi-asserted-by":"crossref","unstructured":"Liu, Z., Yeh, R.A., Tang, X., Liu, Y., Agarwala, A.: Video frame synthesis using deep voxel flow. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4463\u20134471 (2017)","DOI":"10.1109\/ICCV.2017.478"},{"key":"22_CR32","doi-asserted-by":"crossref","unstructured":"Lugmayr, A., Danelljan, M., Romero, A., Yu, F., Timofte, R., Van\u00a0Gool, L.: Repaint: inpainting using denoising diffusion probabilistic models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11461\u201311471 (2022)","DOI":"10.1109\/CVPR52688.2022.01117"},{"key":"22_CR33","doi-asserted-by":"crossref","unstructured":"Luo, Z., et al.: Videofusion: decomposed diffusion models for high-quality video generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10209\u201310218 (2023)","DOI":"10.1109\/CVPR52729.2023.00984"},{"key":"22_CR34","doi-asserted-by":"crossref","unstructured":"Mahapatra, A., Kulkarni, K.: Controllable animation of fluid elements in still images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3667\u20133676 (2022)","DOI":"10.1109\/CVPR52688.2022.00365"},{"issue":"6","key":"22_CR35","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3618326","volume":"42","author":"A Mahapatra","year":"2023","unstructured":"Mahapatra, A., Siarohin, A., Lee, H.Y., Tulyakov, S., Zhu, J.Y.: Text-guided synthesis of eulerian cinemagraphs. ACM Trans. Graph. (TOG) 42(6), 1\u201313 (2023)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"22_CR36","unstructured":"Meng, C., et al.: Sdedit: guided image synthesis and editing with stochastic differential equations. arXiv preprint arXiv:2108.01073 (2021)"},{"key":"22_CR37","doi-asserted-by":"crossref","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: Nerf: representing scenes as neural radiance fields for view synthesis. In: ECCV (2020)","DOI":"10.1007\/978-3-030-58452-8_24"},{"key":"22_CR38","doi-asserted-by":"crossref","unstructured":"Mokady, R., Hertz, A., Aberman, K., Pritch, Y., Cohen-Or, D.: Null-text inversion for editing real images using guided diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6038\u20136047 (2023)","DOI":"10.1109\/CVPR52729.2023.00585"},{"issue":"4","key":"22_CR39","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530127","volume":"41","author":"T M\u00fcller","year":"2022","unstructured":"M\u00fcller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. (ToG) 41(4), 1\u201315 (2022)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"22_CR40","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Barron, J.T., Mildenhall, B., Sajjadi, M.S., Geiger, A., Radwan, N.: Regnerf: regularizing neural radiance fields for view synthesis from sparse inputs. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5480\u20135490 (2022)","DOI":"10.1109\/CVPR52688.2022.00540"},{"key":"22_CR41","doi-asserted-by":"crossref","unstructured":"Niklaus, S., Mai, L., Liu, F.: Video frame interpolation via adaptive convolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 670\u2013679 (2017)","DOI":"10.1109\/CVPR.2017.244"},{"key":"22_CR42","unstructured":"Qiu, Z., et al.: Controlling text-to-image diffusion by orthogonal finetuning. In: Advances in Neural Information Processing Systems, vol. 36 (2024)"},{"key":"22_CR43","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"22_CR44","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"250","DOI":"10.1007\/978-3-031-20071-7_15","volume-title":"ECCV 2022","author":"F Reda","year":"2022","unstructured":"Reda, F., Kontkanen, J., Tabellion, E., Sun, D., Pantofaru, C., Curless, B.: FILM: frame interpolation for large motion. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13667, pp. 250\u2013266. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20071-7_15"},{"key":"22_CR45","doi-asserted-by":"crossref","unstructured":"Sch\u00f6nberger, J.L., Frahm, J.M.: Structure-from-motion revisited. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2016)","DOI":"10.1109\/CVPR.2016.445"},{"key":"22_CR46","doi-asserted-by":"crossref","unstructured":"Sch\u00f6nberger, J.L., Zheng, E., Pollefeys, M., Frahm, J.M.: Pixelwise view selection for unstructured multi-view stereo. In: European Conference on Computer Vision (ECCV) (2016)","DOI":"10.1007\/978-3-319-46487-9_31"},{"key":"22_CR47","unstructured":"Seitzer, M.: Pytorch-fid: FID Score for PyTorch (2020). Version 0.3.0. https:\/\/github.com\/mseitzer\/pytorch-fid"},{"key":"22_CR48","doi-asserted-by":"crossref","unstructured":"Sim, H., Oh, J., Kim, M.: XVFI: extreme video frame interpolation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14489\u201314498 (2021)","DOI":"10.1109\/ICCV48922.2021.01422"},{"key":"22_CR49","unstructured":"Sitzmann, V., Zollh\u00f6fer, M., Wetzstein, G.: Scene representation networks: continuous 3D-structure-aware neural scene representations. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"22_CR50","doi-asserted-by":"crossref","unstructured":"Trevithick, A., Yang, B.: GRF: learning a general radiance field for 3D representation and rendering. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 15182\u201315192 (2021)","DOI":"10.1109\/ICCV48922.2021.01490"},{"key":"22_CR51","doi-asserted-by":"crossref","unstructured":"Tumanyan, N., Geyer, M., Bagon, S., Dekel, T.: Plug-and-play diffusion features for text-driven image-to-image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1921\u20131930 (2023)","DOI":"10.1109\/CVPR52729.2023.00191"},{"key":"22_CR52","unstructured":"Unterthiner, T., van Steenkiste, S., Kurach, K., Marinier, R., Michalski, M., Gelly, S.: FVD: a new metric for video generation (2019)"},{"key":"22_CR53","unstructured":"Wang, X., et al.: Videocomposer: compositional video synthesis with motion controllability. In: Advances in Neural Information Processing Systems, vol. 36 (2024)"},{"key":"22_CR54","unstructured":"Wang, Y., et al.: Lavie: high-quality video generation with cascaded latent diffusion models. arXiv preprint arXiv:2309.15103 (2023)"},{"key":"22_CR55","doi-asserted-by":"crossref","unstructured":"Wang, Z., et al.: Motionctrl: a unified and flexible motion controller for video generation. arXiv preprint arXiv:2312.03641 (2023)","DOI":"10.1145\/3641519.3657518"},{"key":"22_CR56","unstructured":"Xu, X., Siyao, L., Sun, W., Yin, Q., Yang, M.H.: Quadratic video interpolation. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"22_CR57","unstructured":"Yin, S., et al.: Dragnuwa: fine-grained control in video generation by integrating text, image, and trajectory. arXiv preprint arXiv:2308.08089 (2023)"},{"key":"22_CR58","doi-asserted-by":"crossref","unstructured":"Yu, A., Ye, V., Tancik, M., Kanazawa, A.: Pixelnerf: neural radiance fields from one or few images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4578\u20134587 (2021)","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"22_CR59","doi-asserted-by":"crossref","unstructured":"Zhang, D.J., et al.: Show-1: marrying pixel and latent diffusion models for text-to-video generation. arXiv preprint arXiv:2309.15818 (2023)","DOI":"10.1007\/s11263-024-02271-9"},{"key":"22_CR60","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Song, J., Huang, X., Chen, Y., Liu, M.Y.: Diffcollage: parallel generation of large content with diffusion models. arXiv preprint arXiv:2303.17076 (2023)","DOI":"10.1109\/CVPR52729.2023.00982"},{"key":"22_CR61","unstructured":"Zhang, Y., Wei, Y., Jiang, D., Zhang, X., Zuo, W., Tian, Q.: Controlvideo: training-free controllable text-to-video generation. arXiv preprint arXiv:2305.13077 (2023)"},{"key":"22_CR62","doi-asserted-by":"crossref","unstructured":"Zhou, Z., Tulsiani, S.: Sparsefusion: distilling view-conditioned diffusion for 3D reconstruction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12588\u201312597 (2023)","DOI":"10.1109\/CVPR52729.2023.01211"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73229-4_22","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T06:34:55Z","timestamp":1732948495000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73229-4_22"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,25]]},"ISBN":["9783031732287","9783031732294"],"references-count":62,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73229-4_22","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,25]]},"assertion":[{"value":"25 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}