{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,30]],"date-time":"2026-04-30T16:41:18Z","timestamp":1777567278800,"version":"3.51.4"},"publisher-location":"Cham","reference-count":67,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031729942","type":"print"},{"value":"9783031729959","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,24]],"date-time":"2024-11-24T00:00:00Z","timestamp":1732406400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,24]],"date-time":"2024-11-24T00:00:00Z","timestamp":1732406400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72995-9_6","type":"book-chapter","created":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T19:17:21Z","timestamp":1732389441000},"page":"90-107","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Mono-ViFI: A Unified Learning Framework for\u00a0Self-supervised Single and\u00a0Multi-frame Monocular Depth Estimation"],"prefix":"10.1007","author":[{"given":"Jinfeng","family":"Liu","sequence":"first","affiliation":[]},{"given":"Lingtong","family":"Kong","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zerong","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Hong","family":"Gu","sequence":"additional","affiliation":[]},{"given":"Jinwei","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,24]]},"reference":[{"key":"6_CR1","doi-asserted-by":"crossref","unstructured":"Bae, J., Moon, S., Im, S.: Deep digging into the generalization of self-supervised monocular depth estimation. In: AAAI, pp. 187\u2013196 (2023)","DOI":"10.1609\/aaai.v37i1.25090"},{"key":"6_CR2","doi-asserted-by":"crossref","unstructured":"Bangunharcana, A., Magd, A., Kim, K.S.: DualRefine: self-supervised depth and pose estimation through iterative epipolar sampling and refinement toward equilibrium. In: CVPR, pp. 726\u2013738 (2023)","DOI":"10.1109\/CVPR52729.2023.00077"},{"key":"6_CR3","unstructured":"Bello, J.L.G., Kim, M.: Forget about the lidar: self-supervised depth estimators with med probability volumes. In: NeurIPS, pp. 12626\u201312637 (2020)"},{"key":"6_CR4","unstructured":"Bello, J.L.G., Moon, J., Kim, M.: Positional information is all you need: A novel pipeline for self-supervised svde from videos. arXiv:2205.08851 (2022)"},{"key":"6_CR5","unstructured":"Bian, J., et al.: Unsupervised scale-consistent depth and ego-motion learning from monocular video. In: NeurIPS, pp. 35\u201345 (2019)"},{"key":"6_CR6","doi-asserted-by":"crossref","unstructured":"Chan, K.C., Wang, X., Yu, K., Dong, C., Loy, C.C.: Basicvsr: the search for essential components in video super-resolution and beyond. In: CVPR, pp. 4945\u20134954 (2021)","DOI":"10.1109\/CVPR46437.2021.00491"},{"key":"6_CR7","doi-asserted-by":"crossref","unstructured":"Chan, K.C., Zhou, S., Xu, X., Loy, C.C.: BasicVSR++: improving video super-resolution with enhanced propagation and alignment. In: CVPR, pp. 5962\u20135971 (2022)","DOI":"10.1109\/CVPR52688.2022.00588"},{"key":"6_CR8","doi-asserted-by":"crossref","unstructured":"Chen, Y., Liu, S., Wang, X.: Learning continuous image representation with local implicit image function. In: CVPR, pp. 8624\u20138634 (2021)","DOI":"10.1109\/CVPR46437.2021.00852"},{"key":"6_CR9","doi-asserted-by":"crossref","unstructured":"Chen, Y., Schmid, C., Sminchisescu, C.: Self-supervised learning with geometric constraints in monocular video: connecting flow, depth, and camera. In: ICCV, pp. 7063\u20137072 (2019)","DOI":"10.1109\/ICCV.2019.00716"},{"key":"6_CR10","doi-asserted-by":"crossref","unstructured":"Cordts, M., et al.: The cityscapes dataset for semantic urban scene understanding. In: CVPR, pp. 3213\u20133223 (2016)","DOI":"10.1109\/CVPR.2016.350"},{"key":"6_CR11","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: CVPR, pp. 248\u2013255 (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"6_CR12","doi-asserted-by":"crossref","unstructured":"Eigen, D., Fergus, R.: Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture. In: ICCV, pp. 2650\u20132658 (2015)","DOI":"10.1109\/ICCV.2015.304"},{"key":"6_CR13","unstructured":"Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network. In: NeurIPS, pp. 2366\u20132374 (2014)"},{"key":"6_CR14","doi-asserted-by":"crossref","unstructured":"Feng, Z., Yang, L., Jing, L., Wang, H., Tian, Y., Li, B.: Disentangling object motion and occlusion for unsupervised multi-frame monocular depth. In: ECCV, pp. 228\u2013244 (2022)","DOI":"10.1007\/978-3-031-19824-3_14"},{"key":"6_CR15","doi-asserted-by":"crossref","unstructured":"Garg, R., Kumar, B.G.V., Carneiro, G., Reid, I.: Unsupervised CNN for single view depth estimation: Geometry to the rescue. In: ECCV, pp. 740\u2013756 (2016)","DOI":"10.1007\/978-3-319-46484-8_45"},{"issue":"11","key":"6_CR16","first-page":"1231","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: the KITTI dataset. IJRR 32(11), 1231\u20131237 (2013)","journal-title":"IJRR"},{"key":"6_CR17","doi-asserted-by":"crossref","unstructured":"Godard, C., Aodha, O.M., Brostow, G.J.: Unsupervised monocular depth estimation with left-right consistency. In: CVPR, pp. 6602\u20136611 (2017)","DOI":"10.1109\/CVPR.2017.699"},{"key":"6_CR18","doi-asserted-by":"crossref","unstructured":"Godard, C., Aodha, O.M., Firman, M., Brostow, G.J.: Digging into self-supervised monocular depth estimation. In: CVPR, pp. 3828\u20133838 (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"key":"6_CR19","doi-asserted-by":"crossref","unstructured":"Gordon, A., Li, H., Jonschkowski, R., Angelova, A.: Depth from videos in the wild: unsupervised monocular depth learning from unknown cameras. In: ICCV, pp. 8976\u20138985 (2019)","DOI":"10.1109\/ICCV.2019.00907"},{"key":"6_CR20","doi-asserted-by":"crossref","unstructured":"Guizilini, V., Ambrus, R., Chen, D., Zakharov, S., Gaidon, A.: Multi-frame self-supervised depth with transformers. In: CVPR, pp. 160\u2013170 (2022)","DOI":"10.1109\/CVPR52688.2022.00026"},{"key":"6_CR21","doi-asserted-by":"crossref","unstructured":"Guizilini, V., Ambrus, R., Pillai, S., Gaidon, A.: 3D packing for self-supervised monocular depth estimation. In: CVPR, pp. 2482\u20132491 (2020)","DOI":"10.1109\/CVPR42600.2020.00256"},{"key":"6_CR22","doi-asserted-by":"crossref","unstructured":"Han, W., Yin, J., Jin, X., Dai, X., Shen, J.: BRNet: exploring comprehensive features for monocular depth estimation. In: ECCV, pp. 586\u2013602 (2022)","DOI":"10.1007\/978-3-031-19839-7_34"},{"key":"6_CR23","doi-asserted-by":"crossref","unstructured":"Han, W., Yin, J., Shen, J.: Self-supervised monocular depth estimation by direction-aware cumulative convolution network. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00791"},{"key":"6_CR24","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"6_CR25","doi-asserted-by":"crossref","unstructured":"He, M., Hui, L., Bian, Y., Ren, J., Xie, J., Yang, J.: Ra-depth: resolution adaptive self-supervised monocular depth estimation. In: ECCV, pp. 565\u2013581 (2022)","DOI":"10.1007\/978-3-031-19812-0_33"},{"key":"6_CR26","doi-asserted-by":"crossref","unstructured":"Huang, Z., Zhang, T., Heng, W., Shi, B., Zhou, S.: Real-time intermediate flow estimation for video frame interpolation. In: ECCV, pp. 624\u2013642 (2022)","DOI":"10.1007\/978-3-031-19781-9_36"},{"key":"6_CR27","doi-asserted-by":"crossref","unstructured":"Jung, H., Park, E., Yoo, S.: Fine-grained semantics-aware representation enhancement for self-supervised monocular depth estimation. In: ICCV, pp. 12622\u201312632 (2021)","DOI":"10.1109\/ICCV48922.2021.01241"},{"key":"6_CR28","doi-asserted-by":"crossref","unstructured":"Klingner, M., Term\u00f6hlen, J.A., Mikolajczyk, J., Fingscheidt, T.: Self-supervised monocular depth estimation: Solving the dynamic object problem by semantic guidance. In: ECCV, pp. 582\u2013600 (2020)","DOI":"10.1007\/978-3-030-58565-5_35"},{"key":"6_CR29","doi-asserted-by":"crossref","unstructured":"Kong, L., Jiang, B., Luo, D., Chu, W., Huang, X., Tai, Y., Wang, C., Yang, J.: IFRNet: intermediate feature refine network for efficient frame interpolation. In: CVPR, pp. 1959\u20131968 (2022)","DOI":"10.1109\/CVPR52688.2022.00201"},{"key":"6_CR30","doi-asserted-by":"crossref","unstructured":"Kuznietsov, Y., Proesmans, M., Gool, L.V.: CoMoDA: continuous monocular depth adaptation using past experiences. In: WACV, pp. 2906\u20132916 (2021)","DOI":"10.1109\/WACV48630.2021.00295"},{"key":"6_CR31","doi-asserted-by":"crossref","unstructured":"Lee, S., Im, S., Lin, S., Kweon, I.S.: Learning monocular depth in dynamic scenes via instance-aware projection consistency. In: AAAI, pp. 1863\u20131872 (2021)","DOI":"10.1609\/aaai.v35i3.16281"},{"key":"6_CR32","doi-asserted-by":"crossref","unstructured":"Lee, S., Rameau, F., Pan, F., Kweon, I.S.: Attentive and contrastive learning for joint depth and motion field estimation. In: ICCV, pp. 4862\u20134871 (2021)","DOI":"10.1109\/ICCV48922.2021.00482"},{"key":"6_CR33","unstructured":"Li, H., Gordon, A., Zhao, H., Casser, V., Angelova, A.: Unsupervised monocular depth learning in dynamic scenes. In: CoRL, pp. 1908\u20131917 (2020)"},{"key":"6_CR34","doi-asserted-by":"crossref","unstructured":"Li, Z., Zhu, Z.L., Han, L.H., Hou, Q., Guo, C.L., Cheng, M.M.: AMT: all-pairs multi-field transforms for efficient frame interpolation. In: CVPR, pp. 9801\u20139810 (2023)","DOI":"10.1109\/CVPR52729.2023.00945"},{"key":"6_CR35","doi-asserted-by":"crossref","unstructured":"Liu, J., Kong, L., Yang, J.: Designing and searching for lightweight monocular depth network. In: ICONIP, pp. 477\u2013488 (2021)","DOI":"10.1007\/978-3-030-92273-3_39"},{"key":"6_CR36","doi-asserted-by":"crossref","unstructured":"Liu, J., Kong, L., Yang, J.: ATCA: an arc trajectory based model with curvature attention for video frame interpolation. In: ICIP, pp. 1486\u20131490 (2022)","DOI":"10.1109\/ICIP46576.2022.9897981"},{"issue":"1","key":"6_CR37","doi-asserted-by":"publisher","first-page":"763","DOI":"10.1109\/LRA.2023.3337594","volume":"9","author":"J Liu","year":"2024","unstructured":"Liu, J., Kong, L., Yang, J., Liu, W.: Towards better data exploitation in self-supervised monocular depth estimation. IEEE Robot. Autom. Lett. 9(1), 763\u2013770 (2024)","journal-title":"IEEE Robot. Autom. Lett."},{"issue":"71","key":"6_CR38","first-page":"1","volume":"39","author":"X Luo","year":"2020","unstructured":"Luo, X., Huang, J.B., Szeliski, R., Matzen, K., Kopf, J.: Consistent video depth estimation. TOG 39(71), 1\u201313 (2020)","journal-title":"TOG"},{"key":"6_CR39","doi-asserted-by":"crossref","unstructured":"Lyu, X., Liu, L., Wang, M., Kong, X., Liu, L., Liu, Y., Chen, X., Yuan, Y.: HR-depth: high resolution self-supervised monocular depth estimation. In: AAAI, pp. 2294\u20132301 (2021)","DOI":"10.1609\/aaai.v35i3.16329"},{"key":"6_CR40","doi-asserted-by":"crossref","unstructured":"Ma, J., Lei, X., Liu, N., Zhao, X., Pu, S.: Towards comprehensive representation enhancement in semantics-guided self-supervised monocular depth estimation. In: ECCV, pp. 304\u2013321 (2022)","DOI":"10.1007\/978-3-031-19769-7_18"},{"key":"6_CR41","doi-asserted-by":"crossref","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: ECCV, pp. 405\u2013421 (2020)","DOI":"10.1007\/978-3-030-58452-8_24"},{"issue":"4","key":"6_CR42","doi-asserted-by":"publisher","first-page":"6813","DOI":"10.1109\/LRA.2020.3017478","volume":"5","author":"V Patil","year":"2020","unstructured":"Patil, V., Gansbeke, W.V., Dai, D., Gool, L.V.: Don\u2019t forget the past: recurrent depth estimation from monocular video. IEEE Robot. Autom. Lett. 5(4), 6813\u20136820 (2020)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"6_CR43","doi-asserted-by":"crossref","unstructured":"Peng, R., Wang, R., Lai, Y., Tang, L., Cai, Y.: Excavating the potential capacity of self supervised monocular depth estimation. In: ICCV, pp. 15560\u201315569 (2021)","DOI":"10.1109\/ICCV48922.2021.01527"},{"key":"6_CR44","doi-asserted-by":"crossref","unstructured":"Petrovai, A., Nedevschi, S.: Exploiting pseudo labels in a self-supervised learning framework for improved monocular depth estimation. In: CVPR, pp. 1568\u20131578 (2022)","DOI":"10.1109\/CVPR52688.2022.00163"},{"key":"6_CR45","doi-asserted-by":"crossref","unstructured":"Pilzer, A., Lathuili\u00e8re, S., Sebe, N., Ricci, E.: Refine and distill: exploiting cycle-inconsistency and knowledge distillation for unsupervised monocular depth estimation. In: CVPR, pp. 9760\u20139769 (2019)","DOI":"10.1109\/CVPR.2019.01000"},{"key":"6_CR46","doi-asserted-by":"crossref","unstructured":"Poggi, M., Aleotti, F., Tosi, F., Mattoccia, S.: On the uncertainty of self-supervised monocular depth estimation. In: CVPR, pp. 3224\u20133234 (2020)","DOI":"10.1109\/CVPR42600.2020.00329"},{"key":"6_CR47","doi-asserted-by":"crossref","unstructured":"Ranjan, A., et al.: Competitive collaboration: joint unsupervised learning of depth, camera motion, optical flow and motion segmentation. In: CVPR, pp. 12240\u201312249 (2019)","DOI":"10.1109\/CVPR.2019.01252"},{"key":"6_CR48","doi-asserted-by":"crossref","unstructured":"Ren, W., Wang, L., Piao, Y., Zhang, M., Lu, H., Liu, T.: Adaptive co-teaching for unsupervised monocular depth estimation. In: ECCV, pp. 89\u2013105 (2022)","DOI":"10.1007\/978-3-031-19769-7_6"},{"key":"6_CR49","doi-asserted-by":"crossref","unstructured":"Ruhkamp, P., Gao, D., Chen, H., Navab, N., Busam, B.: Attention meets geometry: Geometry guided spatial-temporal attention for consistent self-supervised monocular depth estimation. In: 3DV, pp. 837\u2013847 (2021)","DOI":"10.1109\/3DV53792.2021.00092"},{"issue":"5","key":"6_CR50","doi-asserted-by":"publisher","first-page":"824","DOI":"10.1109\/TPAMI.2008.132","volume":"31","author":"A Saxena","year":"2009","unstructured":"Saxena, A., Sun, M., Ng, A.Y.: Make3D: learning 3D scene structure from a single still image. PAMI 31(5), 824\u2013840 (2009)","journal-title":"PAMI"},{"key":"6_CR51","unstructured":"Tancik, M., et al.: Fourier features let networks learn high frequency functions in low dimensional domains. In: NeurIPS, pp. 7537\u20137547 (2020)"},{"key":"6_CR52","doi-asserted-by":"crossref","unstructured":"Uhrig, J., Schneider, N., Schneider, L., Franke, U., Brox, T., Geiger, A.: Sparsity invariant CNNs. In: 3DV, pp. 11\u201320 (2017)","DOI":"10.1109\/3DV.2017.00012"},{"key":"6_CR53","doi-asserted-by":"crossref","unstructured":"Wang, C., Buenaposada, J.M., Zhu, R., Lucey, S.: Learning depth from monocular videos using direct methods. In: CVPR, pp. 2022\u20132030 (2018)","DOI":"10.1109\/CVPR.2018.00216"},{"key":"6_CR54","doi-asserted-by":"crossref","unstructured":"Wang, R., Pizer, S.M., Frahm, J.M.: Recurrent neural network for (un-)supervised learning of monocular video visual odometry and depth. In: CVPR, pp. 5555\u20135564 (2019)","DOI":"10.1109\/CVPR.2019.00570"},{"key":"6_CR55","doi-asserted-by":"crossref","unstructured":"Wang, R., Yu, Z., Gao, S.: PlaneDepth: self-supervised depth estimation via orthogonal planes. In: CVPR, pp. 21425\u201321434 (2023)","DOI":"10.1109\/CVPR52729.2023.02052"},{"key":"6_CR56","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: Crafting monocular cues and velocity guidance for self-supervised multi-frame depth learning. In: AAAI, pp. 2689\u20132697 (2023)","DOI":"10.1609\/aaai.v37i3.25368"},{"issue":"4","key":"6_CR57","first-page":"600","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE TIP 13(4), 600\u2013612 (2004)","journal-title":"IEEE TIP"},{"key":"6_CR58","doi-asserted-by":"crossref","unstructured":"Watson, J., Aodha, O.M., Prisacariu, V.A., Brostow, G.J., Firman, M.: The temporal opportunist: Self-supervised multi-frame monocular depth. In: CVPR, pp. 1164\u20131174 (2021)","DOI":"10.1109\/CVPR46437.2021.00122"},{"key":"6_CR59","doi-asserted-by":"crossref","unstructured":"Watson, J., Firman, M., Brostow, G.J., Turmukhambetov, D.: Self-supervised monocular depth hints. In: ICCV, pp. 2162\u20132171 (2019)","DOI":"10.1109\/ICCV.2019.00225"},{"key":"6_CR60","doi-asserted-by":"crossref","unstructured":"Yin, Z., Shi, J.: GeoNet: unsupervised learning of dense depth, optical flow and camera pose. In: CVPR, pp. 1983\u20131992 (2018)","DOI":"10.1109\/CVPR.2018.00212"},{"key":"6_CR61","doi-asserted-by":"crossref","unstructured":"Zhang, H., Li, Y., Cao, Y., Liu, Y., Shen, C., Yan, Y.: Exploiting temporal consistency for real-time video depth estimation. In: ICCV, pp. 1725\u20131734 (2019)","DOI":"10.1109\/ICCV.2019.00181"},{"key":"6_CR62","doi-asserted-by":"crossref","unstructured":"Zhang, N., Nex, F., Vosselman, G., Kerle, N.: Lite-Mono: a lightweight CNN and transformer architecture for self-supervised monocular depth estimation. In: CVPR, pp. 18537\u201318546 (2023)","DOI":"10.1109\/CVPR52729.2023.01778"},{"key":"6_CR63","doi-asserted-by":"crossref","unstructured":"Zhao, C., et al.: GasMono: geometry-aided self-supervised monocular depth estimation for indoor scenes. In: ICCV, pp. 16163\u201316174 (2023)","DOI":"10.1109\/ICCV51070.2023.01485"},{"key":"6_CR64","doi-asserted-by":"crossref","unstructured":"Zhao, C., et al.: MonoViT: self-supervised monocular depth estimation with a vision transformer. In: 3DV, pp. 668\u2013678 (2022)","DOI":"10.1109\/3DV57658.2022.00077"},{"key":"6_CR65","unstructured":"Zhou, H., Greenwood, D., Taylor, S.: Self-supervised monocular depth estimation with internal feature fusion. In: BMVC (2021)"},{"key":"6_CR66","doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M., Snavely, N., Lowe, D.G.: Unsupervised learning of depth and ego-motion from video. In: CVPR, pp. 6612\u20136619 (2017)","DOI":"10.1109\/CVPR.2017.700"},{"key":"6_CR67","doi-asserted-by":"crossref","unstructured":"Zhou, Z., Fan, X., Shi, P., Xin, Y.: R-MSFM: recurrent multi-scale feature modulation for monocular depth estimating. In: ICCV, pp. 12757\u201312766 (2021)","DOI":"10.1109\/ICCV48922.2021.01254"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72995-9_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T20:02:42Z","timestamp":1732392162000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72995-9_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,24]]},"ISBN":["9783031729942","9783031729959"],"references-count":67,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72995-9_6","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,24]]},"assertion":[{"value":"24 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}