{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T15:31:40Z","timestamp":1774539100457,"version":"3.50.1"},"publisher-location":"Cham","reference-count":56,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031726323","type":"print"},{"value":"9783031726330","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T00:00:00Z","timestamp":1732233600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T00:00:00Z","timestamp":1732233600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72633-0_16","type":"book-chapter","created":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T07:56:02Z","timestamp":1732175762000},"page":"285-302","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Latent-INR: A Flexible Framework for\u00a0Implicit Representations of\u00a0Videos with\u00a0Discriminative Semantics"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5346-9510","authenticated-orcid":false,"given":"Shishira R.","family":"Maiya","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4017-2862","authenticated-orcid":false,"given":"Anubhav","family":"Gupta","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9826-6285","authenticated-orcid":false,"given":"Matthew","family":"Gwilliam","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3975-5788","authenticated-orcid":false,"given":"Max","family":"Ehrlich","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8928-8554","authenticated-orcid":false,"given":"Abhinav","family":"Shrivastava","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,22]]},"reference":[{"key":"16_CR1","unstructured":"A scheme for shot detection and video retreival using spatio temporal features. Int. J. Recent Technol. Eng. (2019). https:\/\/api.semanticscholar.org\/CorpusID:241499090"},{"key":"16_CR2","doi-asserted-by":"crossref","unstructured":"Bai, Y., Dong, C., Wang, C., Yuan, C.: Ps-NeRV: patch-wise stylized neural representations for videos. In: 2023 IEEE International Conference on Image Processing (ICIP), pp. 41\u201345. IEEE (2023)","DOI":"10.1109\/ICIP49359.2023.10222144"},{"key":"16_CR3","doi-asserted-by":"crossref","unstructured":"Bain, M., Nagrani, A., Varol, G., Zisserman, A.: Frozen in time: a joint video and image encoder for end-to-end retrieval. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 1708\u20131718 (2021). https:\/\/api.semanticscholar.org\/CorpusID:232478955","DOI":"10.1109\/ICCV48922.2021.00175"},{"key":"16_CR4","unstructured":"Ball\u00e9, J., Minnen, D.C., Singh, S., Hwang, S.J., Johnston, N.: Variational image compression with a scale hyperprior. ArXiv abs\/1802.01436 (2018). https:\/\/api.semanticscholar.org\/CorpusID:3611540"},{"key":"16_CR5","unstructured":"Bauer, M., Dupont, E., Brock, A., Rosenbaum, D., Schwarz, J., Kim, H.: Spatial functa: scaling functa to ImageNet classification and generation. arXiv preprint arXiv:2302.03130 (2023)"},{"key":"16_CR6","doi-asserted-by":"crossref","unstructured":"Buch, S., Eyzaguirre, C., Gaidon, A., Wu, J., Fei-Fei, L., Niebles, J.C.: Revisiting the \u201cvideo\u201d in video-language understanding (2022)","DOI":"10.1109\/CVPR52688.2022.00293"},{"key":"16_CR7","doi-asserted-by":"publisher","unstructured":"Cao, L., Jiang, A., Li, W., Wu, H., Ye, N.: OoDHDR-codec: Out-of-distribution generalization for HDR image compression. In: Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, 22 February\u20131 March 2022, pp. 158\u2013166. AAAI Press (2022). https:\/\/doi.org\/10.1609\/AAAI.V36I1.19890","DOI":"10.1609\/AAAI.V36I1.19890"},{"key":"16_CR8","doi-asserted-by":"crossref","unstructured":"Chen, H., Gwilliam, M., Lim, S.N., Shrivastava, A.: Hnerv: A hybrid neural representation for videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10270\u201310279 (2023)","DOI":"10.1109\/CVPR52729.2023.00990"},{"key":"16_CR9","first-page":"21557","volume":"34","author":"H Chen","year":"2021","unstructured":"Chen, H., He, B., Wang, H., Ren, Y., Lim, S.N., Shrivastava, A.: NeRV: neural representations for videos. Adv. Neural. Inf. Process. Syst. 34, 21557\u201321568 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"16_CR10","unstructured":"Chen, H., Matthew, G., He, B., Lim, S.N., Shrivastava, A.: CNeRV: content-adaptive neural representation for visual data. In: BMVC (2022)"},{"key":"16_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"170","DOI":"10.1007\/978-3-031-19790-1_11","volume-title":"Computer Vision \u2013 ECCV 2022","author":"Y Chen","year":"2022","unstructured":"Chen, Y., Wang, X.: Transformers as meta-learners for implicit neural representations. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13677, pp. 170\u2013187. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19790-1_11"},{"key":"16_CR12","doi-asserted-by":"crossref","unstructured":"Chen, Y., et\u00a0al.: An overview of core coding tools in the AV1 video codec. In: 2018 picture coding symposium (PCS), pp. 41\u201345. IEEE (2018)","DOI":"10.1109\/PCS.2018.8456249"},{"key":"16_CR13","doi-asserted-by":"crossref","unstructured":"Chiang, P.Z., Tsai, M.S., Tseng, H.Y., Lai, W.S., Chiu, W.C.: Stylizing 3D scene via implicit representation and hypernetwork. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1475\u20131484 (2022)","DOI":"10.1109\/WACV51458.2022.00029"},{"key":"16_CR14","unstructured":"Dupont, E., Kim, H., Eslami, S.M.A., Rezende, D.J., Rosenbaum, D.: From data to functa: your data point is a function and you can treat it like one. In: International Conference on Machine Learning (2022). https:\/\/api.semanticscholar.org\/CorpusID:249395684"},{"key":"16_CR15","doi-asserted-by":"crossref","unstructured":"Ehrlich, M., Davis, L.S.: Deep residual learning in the jpeg transform domain. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00358"},{"key":"16_CR16","unstructured":"Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: International Conference on Machine Learning (2017). https:\/\/api.semanticscholar.org\/CorpusID:6719686"},{"key":"16_CR17","unstructured":"Frankle, J., Carbin, M.: The lottery ticket hypothesis: finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635 (2018)"},{"key":"16_CR18","doi-asserted-by":"crossref","unstructured":"He, B., et al.: Towards scalable neural representation for diverse videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6132\u20136142 (2023)","DOI":"10.1109\/CVPR52729.2023.00594"},{"key":"16_CR19","doi-asserted-by":"crossref","unstructured":"Jiang, H., Sun, D., Jampani, V., Yang, M.H., Learned-Miller, E.G., Kautz, J.: Super SloMo: high quality estimation of multiple intermediate frames for video interpolation. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9000\u20139008 (2017). https:\/\/api.semanticscholar.org\/CorpusID:10817557","DOI":"10.1109\/CVPR.2018.00938"},{"key":"16_CR20","doi-asserted-by":"crossref","unstructured":"Kim, C., Lee, D., Kim, S., Cho, M., Han, W.S.: Generalizable implicit neural representations via instance pattern composers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11808\u201311817 (2023)","DOI":"10.1109\/CVPR52729.2023.01136"},{"key":"16_CR21","first-page":"12718","volume":"35","author":"S Kim","year":"2022","unstructured":"Kim, S., Yu, S., Lee, J., Shin, J.: Scalable neural video representations with learnable positional features. Adv. Neural. Inf. Process. Syst. 35, 12718\u201312731 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"16_CR22","doi-asserted-by":"crossref","unstructured":"Krishna, R., Hata, K., Ren, F., Fei-Fei, L., Niebles, J.C.: Dense-captioning events in videos (2017)","DOI":"10.1109\/ICCV.2017.83"},{"key":"16_CR23","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"267","DOI":"10.1007\/978-3-031-19833-5_16","volume-title":"Computer Vision \u2013 ECCV 2022","author":"Z Li","year":"2022","unstructured":"Li, Z., Wang, M., Pi, H., Xu, K., Mei, J., Liu, Y.: E-NeRV: expedite neural video representation with disentangled spatial-temporal context. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13695, pp. 267\u2013284. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19833-5_16"},{"key":"16_CR24","doi-asserted-by":"crossref","unstructured":"Luo, H., et al.: CLIP4Clip: an empirical study of clip for end to end video CLIP retrieval. Neurocomputing 508, 293\u2013304 (2021). https:\/\/api.semanticscholar.org\/CorpusID:233296206","DOI":"10.1016\/j.neucom.2022.07.028"},{"key":"16_CR25","doi-asserted-by":"crossref","unstructured":"Luo, H., Chen, Y., Zhou, Y.: An extremely effective spatial pyramid and pixel shuffle upsampling decoder for multiscale monocular depth estimation. Comput. Intell. Neurosci. 2022 (2022). https:\/\/api.semanticscholar.org\/CorpusID:251272212","DOI":"10.1155\/2022\/4668001"},{"key":"16_CR26","doi-asserted-by":"crossref","unstructured":"Maiya, S.R., et al.: NIRVANA: neural implicit representations of videos with adaptive networks and autoregressive patch-wise modeling. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14378\u201314387 (2023)","DOI":"10.1109\/CVPR52729.2023.01382"},{"key":"16_CR27","unstructured":"McInnes, L., Healy, J., Melville, J.: UMAP: uniform manifold approximation and projection for dimension reduction (2020)"},{"key":"16_CR28","unstructured":"Mentzer, F., Toderici, G., Tschannen, M., Agustsson, E.: High-fidelity generative image compression. ArXiv abs\/2006.09965 (2020). https:\/\/api.semanticscholar.org\/CorpusID:219721015"},{"key":"16_CR29","doi-asserted-by":"publisher","unstructured":"Mercat, A., Viitanen, M., Vanne, J.: UVG dataset: 50\/120fps 4K sequences for video codec analysis and development. In: Proceedings of the 11th ACM Multimedia Systems Conference, MMSys 2020, pp. 297\u2013302. Association for Computing Machinery, New York (2020). https:\/\/doi.org\/10.1145\/3339825.3394937","DOI":"10.1145\/3339825.3394937"},{"key":"16_CR30","doi-asserted-by":"publisher","unstructured":"M\u00fcller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. 41(4), 102:1\u2013102:15 (2022). https:\/\/doi.org\/10.1145\/3528223.3530127","DOI":"10.1145\/3528223.3530127"},{"key":"16_CR31","doi-asserted-by":"crossref","unstructured":"Padmanabhan, N., Gwilliam, M., Kumar, P., Maiya, S.R., Ehrlich, M., Shrivastava, A.: Explaining the implicit neural canvas: connecting pixels to neurons by tracing their contributions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10957\u201310967 (2024)","DOI":"10.1109\/CVPR52733.2024.01042"},{"key":"16_CR32","doi-asserted-by":"crossref","unstructured":"Papa, S., et al.: How to train neural field representations: a comprehensive study and benchmark. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 22616\u201322625 (2024)","DOI":"10.1109\/CVPR52733.2024.02134"},{"key":"16_CR33","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning (2021). https:\/\/api.semanticscholar.org\/CorpusID:231591445"},{"key":"16_CR34","doi-asserted-by":"crossref","unstructured":"Ramanujan, V., Wortsman, M., Kembhavi, A., Farhadi, A., Rastegari, M.: What\u2019s hidden in a randomly weighted neural network? In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 11893\u201311902 (2020)","DOI":"10.1109\/CVPR42600.2020.01191"},{"key":"16_CR35","doi-asserted-by":"crossref","unstructured":"Saragadam, V., LeJeune, D., Tan, J., Balakrishnan, G., Veeraraghavan, A., Baraniuk, R.: WIRE: wavelet implicit neural representations. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) pp. 18507\u201318516 (2023). https:\/\/api.semanticscholar.org\/CorpusID:255749557","DOI":"10.1109\/CVPR52729.2023.01775"},{"key":"16_CR36","unstructured":"Saragadam, V., Tan, J., Balakrishnan, G., Baraniuk, R.G., Veeraraghavan, A.: MINER: multiscale implicit neural representations. CoRR abs\/2202.03532 (2022), https:\/\/arxiv.org\/abs\/2202.03532"},{"key":"16_CR37","unstructured":"Schwarz, J.R., Tack, J., Teh, Y.W., Lee, J., Shin, J.: Modality-agnostic variational compression of implicit neural representations. arXiv preprint arXiv:2301.09479 (2023)"},{"key":"16_CR38","unstructured":"Schwarz, J.R., Teh, Y.W.: Meta-learning sparse compression networks (2022)"},{"key":"16_CR39","unstructured":"Sen, B., Agarwal, A., Namboodiri, V.P., Jawahar, C.: INR-V: a continuous representation space for video-based generative tasks. arXiv preprint arXiv:2210.16579 (2022)"},{"key":"16_CR40","unstructured":"Sen, B., Singh, G., Agarwal, A., Agaram, R., Krishna, K.M., Sridhar, S.: Hyp-NeRF: learning improved nerf priors using a hypernetwork. arXiv preprint arXiv:2306.06093 (2023)"},{"key":"16_CR41","doi-asserted-by":"crossref","unstructured":"Shi, W., et al.: Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1874\u20131883 (2016). https:\/\/api.semanticscholar.org\/CorpusID:7037846","DOI":"10.1109\/CVPR.2016.207"},{"key":"16_CR42","unstructured":"Sitzmann, V., Martel, J.N.P., Bergman, A.W., Lindell, D.B., Wetzstein, G.: Implicit neural representations with periodic activation functions (2020)"},{"key":"16_CR43","first-page":"19313","volume":"34","author":"V Sitzmann","year":"2021","unstructured":"Sitzmann, V., Rezchikov, S., Freeman, B., Tenenbaum, J., Durand, F.: Light field networks: neural scene representations with single-evaluation rendering. Adv. Neural. Inf. Process. Syst. 34, 19313\u201319325 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"16_CR44","unstructured":"Sitzmann, V., Zollh\u00f6fer, M., Wetzstein, G.: Scene representation networks: continuous 3D-structure-aware neural scene representations. Adv. Neural. Inf. Process. Syst. 32 (2019)"},{"key":"16_CR45","doi-asserted-by":"crossref","unstructured":"Skorokhodov, I., Ignatyev, S., Elhoseiny, M.: Adversarial generation of continuous images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10753\u201310764 (2021)","DOI":"10.1109\/CVPR46437.2021.01061"},{"key":"16_CR46","doi-asserted-by":"publisher","unstructured":"Str\u00fcmpler, Y., Postels, J., Yang, R., Gool, L.V., Tombari, F.: Implicit neural representations for image compression. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds) ECCV 2022. LNCS, vol. 13686, pp. 74\u201391. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-031-19809-0_5, https:\/\/api.semanticscholar.org\/CorpusID:244954443","DOI":"10.1007\/978-3-031-19809-0_5"},{"issue":"12","key":"16_CR47","doi-asserted-by":"publisher","first-page":"1649","DOI":"10.1109\/TCSVT.2012.2221191","volume":"22","author":"GJ Sullivan","year":"2012","unstructured":"Sullivan, G.J., Ohm, J.R., Han, W.J., Wiegand, T.: Overview of the high efficiency video coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 22(12), 1649\u20131668 (2012)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"16_CR48","unstructured":"Tack, J., Kim, S., Yu, S., Lee, J., Shin, J., Schwarz, J.R.: Learning large-scale neural fields via context pruned meta-learning (2023)"},{"key":"16_CR49","doi-asserted-by":"crossref","unstructured":"Tancik, M., et al.: Learned initializations for optimizing coordinate-based neural representations. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00287"},{"key":"16_CR50","unstructured":"Tancik, M., et al.: Fourier features let networks learn high frequency functions in low dimensional domains. In: NeurIPS (2020)"},{"key":"16_CR51","doi-asserted-by":"crossref","unstructured":"Tang, Y., et al.: COIN: a large-scale dataset for comprehensive instructional video analysis (2019)","DOI":"10.1109\/CVPR.2019.00130"},{"issue":"4","key":"16_CR52","doi-asserted-by":"publisher","first-page":"30","DOI":"10.1145\/103085.103089","volume":"34","author":"GK Wallace","year":"1991","unstructured":"Wallace, G.K.: The jpeg still picture compression standard. Commun. ACM 34(4), 30\u201344 (1991)","journal-title":"Commun. ACM"},{"key":"16_CR53","doi-asserted-by":"crossref","unstructured":"Wiegand, T., Sullivan, G., Bjontegaard, G., Luthra, A.: Overview of the H.264\/AVC video coding standard. IEEE Trans. Circuits Syst. Video Technol. (2003)","DOI":"10.1109\/TCSVT.2003.815165"},{"key":"16_CR54","doi-asserted-by":"crossref","unstructured":"Xu, J., Mei, T., Yao, T., Rui, Y.: MSR-VTT: a large video description dataset for bridging video and language. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (June 2016)","DOI":"10.1109\/CVPR.2016.571"},{"key":"16_CR55","doi-asserted-by":"crossref","unstructured":"Zhang, H., Li, X., Bing, L.: Video-LLaMA: an instruction-tuned audio-visual language model for video understanding. arXiv preprint arXiv:2306.02858 (2023). https:\/\/arxiv.org\/abs\/2306.02858","DOI":"10.18653\/v1\/2023.emnlp-demo.49"},{"key":"16_CR56","unstructured":"Zhang, M., Zhang, A., McDonagh, S.G.: On the out-of-distribution generalization of probabilistic image modelling. In: Neural Information Processing Systems (2021). https:\/\/api.semanticscholar.org\/CorpusID:237431305"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72633-0_16","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T08:13:22Z","timestamp":1732176802000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72633-0_16"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,22]]},"ISBN":["9783031726323","9783031726330"],"references-count":56,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72633-0_16","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,22]]},"assertion":[{"value":"22 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}