{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,12]],"date-time":"2026-04-12T06:32:30Z","timestamp":1775975550278,"version":"3.50.1"},"publisher-location":"Cham","reference-count":65,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031730320","type":"print"},{"value":"9783031730337","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T00:00:00Z","timestamp":1730332800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T00:00:00Z","timestamp":1730332800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73033-7_2","type":"book-chapter","created":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T00:03:55Z","timestamp":1730333035000},"page":"18-35","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":140,"title":["CoTracker: It Is Better to\u00a0Track Together"],"prefix":"10.1007","author":[{"given":"Nikita","family":"Karaev","sequence":"first","affiliation":[]},{"given":"Ignacio","family":"Rocco","sequence":"additional","affiliation":[]},{"given":"Benjamin","family":"Graham","sequence":"additional","affiliation":[]},{"given":"Natalia","family":"Neverova","sequence":"additional","affiliation":[]},{"given":"Andrea","family":"Vedaldi","sequence":"additional","affiliation":[]},{"given":"Christian","family":"Rupprecht","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,31]]},"reference":[{"issue":"10","key":"2_CR1","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1145\/2001269.2001293","volume":"54","author":"S Agarwal","year":"2011","unstructured":"Agarwal, S., et al.: Building Rome in a day. Commun. ACM 54(10), 105\u2013112 (2011)","journal-title":"Commun. ACM"},{"key":"2_CR2","unstructured":"Bertasius, G., Wang, H., Torresani, L.: Is space-time attention all you need for video understanding? In: Proceedings of the ICML (2021)"},{"key":"2_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"850","DOI":"10.1007\/978-3-319-48881-3_56","volume-title":"Computer Vision \u2013 ECCV 2016 Workshops","author":"L Bertinetto","year":"2016","unstructured":"Bertinetto, L., Valmadre, J., Henriques, J.F., Vedaldi, A., Torr, P.H.S.: Fully-convolutional siamese networks for object tracking. In: Hua, G., J\u00e9gou, H. (eds.) ECCV 2016. LNCS, vol. 9914, pp. 850\u2013865. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-48881-3_56"},{"key":"2_CR4","doi-asserted-by":"crossref","unstructured":"Bhat, G., Danelljan, M., Gool, L.V., Timofte, R.: Learning discriminative model prediction for tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00628"},{"key":"2_CR5","doi-asserted-by":"crossref","unstructured":"Birchfield, S.T., Pundlik, S.J.: Joint tracking of features and edges. In: Proceedings of the CVPR (2008)","DOI":"10.1109\/CVPR.2008.4587486"},{"key":"2_CR6","unstructured":"Black, M.J., Anandan, P.: A framework for the robust estimation of optical flow. In: Proceedings of the ICCV (1993)"},{"key":"2_CR7","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1023\/B:VISI.0000045324.43199.43","volume":"61","author":"A Bruhn","year":"2005","unstructured":"Bruhn, A., Weickert, J., Schn\u00f6rr, C.: Lucas\/kanade meets horn\/schunck: combining local and global optic flow methods. Int. J. Comput. Vis. 61, 211\u2013231 (2005)","journal-title":"Int. J. Comput. Vis."},{"key":"2_CR8","doi-asserted-by":"crossref","unstructured":"Butler, D.J., Wulff, J., Stanley, G.B., Black, M.J.: A naturalistic open source movie for optical flow evaluation. In: Proceedings of the ECCV (2012)","DOI":"10.1007\/978-3-642-33783-3_44"},{"key":"2_CR9","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Proceedings of the ECCV (2020)","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"2_CR10","doi-asserted-by":"publisher","first-page":"103508","DOI":"10.1016\/j.cviu.2022.103508","volume":"222","author":"F Chen","year":"2022","unstructured":"Chen, F., Wang, X., Zhao, Y., Lv, S., Niu, X.: Visual object tracking: a survey. Comput. Vis. Image Underst. 222, 103508 (2022)","journal-title":"Comput. Vis. Image Underst."},{"key":"2_CR11","doi-asserted-by":"crossref","unstructured":"Chen, X., Yan, B., Zhu, J., Wang, D., Yang, X., Lu, H.: Transformer tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8126\u20138135 (2021)","DOI":"10.1109\/CVPR46437.2021.00803"},{"issue":"6","key":"2_CR12","doi-asserted-by":"publisher","first-page":"773","DOI":"10.1109\/83.336247","volume":"3","author":"TM Chin","year":"1994","unstructured":"Chin, T.M., Karl, W.C., Willsky, A.S.: Probabilistic and sequential computation of optical flow using temporal coherence. IEEE Trans. on Image Process. 3(6), 773\u2013788 (1994)","journal-title":"IEEE Trans. on Image Process."},{"key":"2_CR13","doi-asserted-by":"crossref","unstructured":"Cui, Y., Jiang, C., Wang, L., Wu, G.: Mixformer: end-to-end tracking with iterative mixed attention. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13608\u201313618 (2022)","DOI":"10.1109\/CVPR52688.2022.01324"},{"key":"2_CR14","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Bhat, G., Khan, F.S., Felsberg, M.: Atom: accurate tracking by overlap maximization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4660\u20134669 (2019)","DOI":"10.1109\/CVPR.2019.00479"},{"key":"2_CR15","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Bhat, G., Shahbaz\u00a0Khan, F., Felsberg, M.: Eco: efficient convolution operators for tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6638\u20136646 (2017)","DOI":"10.1109\/CVPR.2017.733"},{"key":"2_CR16","unstructured":"Darcet, T., Oquab, M., Mairal, J., Bojanowski, P.: Vision transformers need registers. arXiv preprint arXiv:2309.16588 (2023)"},{"key":"2_CR17","unstructured":"Doersch, C., et al.: TAP-vid: a benchmark for tracking any point in a video. arXiv arXiv:2211.03726 (2022)"},{"key":"2_CR18","doi-asserted-by":"crossref","unstructured":"Doersch, C., et al.: Tapir: tracking any point with per-frame initialization and temporal refinement. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10061\u201310072 (2023)","DOI":"10.1109\/ICCV51070.2023.00923"},{"key":"2_CR19","doi-asserted-by":"crossref","unstructured":"Dosovitskiy, A., et al.: FlowNet: learning optical flow with convolutional networks. In: Proceedings of the ICCV (2015)","DOI":"10.1109\/ICCV.2015.316"},{"issue":"2","key":"2_CR20","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1006\/jvci.1998.0382","volume":"9","author":"M Elad","year":"1998","unstructured":"Elad, M., Feuer, A.: Recursive optical flow estimation\u2013adaptive filtering approach. J. Vis. Commun. Image Represent. 9(2), 119\u2013138 (1998)","journal-title":"J. Vis. Commun. Image Represent."},{"key":"2_CR21","doi-asserted-by":"crossref","unstructured":"Girshick, R., Iandola, F., Darrell, T., Malik, J.: Deformable part models are convolutional neural networks. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp. 437\u2013446 (2015)","DOI":"10.1109\/CVPR.2015.7298641"},{"key":"2_CR22","unstructured":"Greff, K., et\u00a0al.: Kubric: a scalable dataset generator. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3749\u20133761 (2022)"},{"key":"2_CR23","doi-asserted-by":"crossref","unstructured":"Harley, A.W., Fang, Z., Fragkiadaki, K.: Particle video revisited: tracking through occlusions using point trajectories. In: Proceedings of the ECCV (2022)","DOI":"10.1007\/978-3-031-20047-2_4"},{"key":"2_CR24","doi-asserted-by":"crossref","unstructured":"Harley, A.W., Fang, Z., Fragkiadaki, K.: Particle videos revisited: tracking through occlusions using point trajectories. In: Proceedings of the ECCV (2022)","DOI":"10.1007\/978-3-031-20047-2_4"},{"key":"2_CR25","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"749","DOI":"10.1007\/978-3-319-46448-0_45","volume-title":"Computer Vision \u2013 ECCV 2016","author":"D Held","year":"2016","unstructured":"Held, D., Thrun, S., Savarese, S.: Learning to track at 100 FPS with deep regression networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 749\u2013765. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_45"},{"issue":"1\u20133","key":"2_CR26","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1016\/0004-3702(81)90024-2","volume":"17","author":"BK Horn","year":"1981","unstructured":"Horn, B.K., Schunck, B.G.: Determining optical flow. Artif. Intell. 17(1\u20133), 185\u2013203 (1981)","journal-title":"Artif. Intell."},{"key":"2_CR27","doi-asserted-by":"crossref","unstructured":"Huang, Z., et al.: FlowFormer: a transformer architecture for optical flow. In: Proceedings of the ECCV (2022)","DOI":"10.1007\/978-3-031-19790-1_40"},{"key":"2_CR28","doi-asserted-by":"crossref","unstructured":"Ilg, E., Mayer, N., Saikia, T., Keuper, M., Dosovitskiy, A., Brox, T.: FlowNet 2.0: evolution of optical flow estimation with deep networks. In: Proceedings of the CVPR (2017)","DOI":"10.1109\/CVPR.2017.179"},{"key":"2_CR29","unstructured":"Jaegle, A., et al.: Perceiver IO: A general architecture for structured inputs & outputs. In: Proceedings of the ICLR (2022)"},{"key":"2_CR30","doi-asserted-by":"crossref","unstructured":"Janai, J., Guney, F., Ranjan, A., Black, M., Geiger, A.: Unsupervised learning of multi-frame optical flow with occlusions. In: Proceedings of the ECCV (2018)","DOI":"10.1007\/978-3-030-01270-0_42"},{"key":"2_CR31","doi-asserted-by":"crossref","unstructured":"Jia, X., Lu, H., Yang, M.H.: Visual tracking via adaptive structural local sparse appearance model. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1822\u20131829. IEEE (2012)","DOI":"10.1109\/CVPR.2012.6247880"},{"key":"2_CR32","doi-asserted-by":"crossref","unstructured":"Jiang, H., Sun, D., Jampani, V., Yang, M.H., Learned-Miller, E., Kautz, J.: Super SloMo: high quality estimation of multiple intermediate frames for video interpolation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9000\u20139008 (2018)","DOI":"10.1109\/CVPR.2018.00938"},{"key":"2_CR33","doi-asserted-by":"crossref","unstructured":"Jiang, S., Lu, Y., Li, H., Hartley, R.: Learning optical flow from a few matches. In: Proceedings of the CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01632"},{"key":"2_CR34","doi-asserted-by":"crossref","unstructured":"Karaev, N., Rocco, I., Graham, B., Neverova, N., Vedaldi, A., Rupprecht, C.: DynamicStereo: consistent dynamic depth from stereo videos. In: Proceedings of the CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01271"},{"key":"2_CR35","doi-asserted-by":"crossref","unstructured":"Li, F., Tian, C., Zuo, W., Zhang, L., Yang, M.H.: Learning spatial-temporal regularized correlation filters for visual tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4904\u20134913 (2018)","DOI":"10.1109\/CVPR.2018.00515"},{"key":"2_CR36","doi-asserted-by":"crossref","unstructured":"Li, Y., Zhu, J., Hoi, S.C.: Reliable patch trackers: robust visual tracking by exploiting reliable patches. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 353\u2013361 (2015)","DOI":"10.1109\/CVPR.2015.7298632"},{"key":"2_CR37","unstructured":"Lucas, B.D., Kanade, T.: An iterative image registration technique with an application to stereo vision. In: Proceedings of the IJCAI, vol.\u00a02 (1981)"},{"key":"2_CR38","unstructured":"Luo, J., Wan, Z., Li, B., Dai, Y.: Continuous parametric optical flow. In: Thirty-Seventh Conference on Neural Information Processing Systems (2023)"},{"issue":"6","key":"2_CR39","doi-asserted-by":"publisher","first-page":"810","DOI":"10.1109\/TPAMI.2004.16","volume":"26","author":"L Matthews","year":"2004","unstructured":"Matthews, L., Ishikawa, T., Baker, S.: The template update problem. IEEE Trans. Pattern Anal. Mach. Intell. 26(6), 810\u2013815 (2004)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2_CR40","doi-asserted-by":"crossref","unstructured":"Mayer, N., et al.: A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: Proceedings of the CVPR (2016)","DOI":"10.1109\/CVPR.2016.438"},{"key":"2_CR41","unstructured":"Nagrani, A., Yang, S., Arnab, A., Jansen, A., Schmid, C., Sun, C.: Attention bottlenecks for multimodal fusion. In: Advance of Neural Information Processing Systems , vol. 34, pp. 14200\u201314213 (2021)"},{"key":"2_CR42","doi-asserted-by":"crossref","unstructured":"Neoral, M., \u015aer\u00fdch, J., Matas, J.: MFT: long-term tracking of every pixel. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 6837\u20136847 (2023)","DOI":"10.1109\/WACV57701.2024.00669"},{"key":"2_CR43","unstructured":"Pont-Tuset, J., Perazzi, F., Caelles, S., Arbel\u00e1ez, P., Sorkine-Hornung, A., Van\u00a0Gool, L.: The 2017 DAVIS challenge on video object segmentation. arXiv arXiv:1704.00675 (2017)"},{"key":"2_CR44","doi-asserted-by":"crossref","unstructured":"Ren, Z., Gallo, O., Sun, D., Yang, M.H., Sudderth, E.B., Kautz, J.: A fusion approach for multi-frame optical flow estimation. In: Proceedings of the WACV (2019)","DOI":"10.1007\/978-3-030-11024-6_53"},{"key":"2_CR45","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1007\/s11263-008-0136-6","volume":"80","author":"P Sand","year":"2008","unstructured":"Sand, P., Teller, S.: Particle video: long-range motion estimation using point trajectories. Int. J. Comput. Vis. 80, 72\u201391 (2008)","journal-title":"Int. J. Comput. Vis."},{"key":"2_CR46","doi-asserted-by":"crossref","unstructured":"Shi, X., et al.: VideoFlow: exploiting temporal cues for multi-frame optical flow estimation. arXiv arXiv:2303.08340 (2023)","DOI":"10.1109\/ICCV51070.2023.01146"},{"key":"2_CR47","doi-asserted-by":"crossref","unstructured":"Shi, X., et al.: FlowFormer++: masked cost volume autoencoding for pretraining optical flow estimation. arXiv arXiv:2303.01237 (2023)","DOI":"10.1109\/CVPR52729.2023.00160"},{"key":"2_CR48","doi-asserted-by":"crossref","unstructured":"Sidenbladh, H., Black, M.J., Fleet, D.J.: Stochastic tracking of 3D human figures using 2D image motion. In: Proceedings of the ECCV (2000)","DOI":"10.1007\/3-540-45053-X_45"},{"key":"2_CR49","doi-asserted-by":"crossref","unstructured":"Song, Y., Ma, C., Gong, L., Zhang, J., Lau, R.W., Yang, M.H.: Crest: convolutional residual learning for visual tracking. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2555\u20132564 (2017)","DOI":"10.1109\/ICCV.2017.279"},{"key":"2_CR50","doi-asserted-by":"crossref","unstructured":"Sui, X., et al.: Craft: cross-attentional flow transformer for robust optical flow. In: Proceedings of the CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01708"},{"key":"2_CR51","doi-asserted-by":"crossref","unstructured":"Sun, D., et al.: AutoFlow: learning a better training set for optical flow. In: Proceedings of the CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00996"},{"key":"2_CR52","doi-asserted-by":"crossref","unstructured":"Sun, D., Yang, X., Liu, M.Y., Kautz, J.: PWC-Net: CNNs for optical flow using pyramid, warping, and cost volume. In: Proceedings of the CVPR (2018)","DOI":"10.1109\/CVPR.2018.00931"},{"key":"2_CR53","unstructured":"Sun, S., Chen, Y., Zhu, Y., Guo, G., Li, G.: SKFlow: learning optical flow with super kernels. arXiv arXiv:2205.14623 (2022)"},{"key":"2_CR54","doi-asserted-by":"crossref","unstructured":"Teed, Z., Deng, J.: Raft: recurrent all-pairs field transforms for optical flow. In: Proceedings of the ECCV (2020)","DOI":"10.1007\/978-3-030-58536-5_24"},{"key":"2_CR55","unstructured":"Teed, Z., Deng, J.: DROID-SLAM: deep visual slam for monocular, stereo, and RGB-D cameras. In: Advances in Neural Information Processing Systems, vol. 34, pp. 16558\u201316569 (2021)"},{"key":"2_CR56","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems (2017)"},{"key":"2_CR57","unstructured":"Wang, J., Zhong, Y., Dai, Y., Zhang, K., Ji, P., Li, H.: Displacement-invariant matching cost learning for accurate optical flow estimation. In: Advances in Neural Information Processing Systems, vol. 33 (2020)"},{"key":"2_CR58","doi-asserted-by":"crossref","unstructured":"Wang, Q., et al.: Tracking everything everywhere all at once. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 19795\u201319806 (2023)","DOI":"10.1109\/ICCV51070.2023.01813"},{"key":"2_CR59","doi-asserted-by":"crossref","unstructured":"Xu, H., Yang, J., Cai, J., Zhang, J., Tong, X.: High-resolution optical flow from 1D attention and correlation. In: Proceedings of the CVPR (2021)","DOI":"10.1109\/ICCV48922.2021.01033"},{"key":"2_CR60","doi-asserted-by":"crossref","unstructured":"Xu, J., Ranftl, R., Koltun, V.: Accurate optical flow via direct cost volume processing. In: Proceedings of the CVPR (2017)","DOI":"10.1109\/CVPR.2017.615"},{"key":"2_CR61","doi-asserted-by":"crossref","unstructured":"Yang, G., Vo, M., Neverova, N., Ramanan, D., Vedaldi, A., Joo, H.: Banmo: building animatable 3D neural models from many casual videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2863\u20132873 (2022)","DOI":"10.1109\/CVPR52688.2022.00288"},{"key":"2_CR62","doi-asserted-by":"crossref","unstructured":"Zhang, F., Woodford, O.J., Prisacariu, V.A., Torr, P.H.: Separable flow: learning motion cost volumes for optical flow estimation. In: Proceedings of the CVPR (2021)","DOI":"10.1109\/ICCV48922.2021.01063"},{"key":"2_CR63","doi-asserted-by":"crossref","unstructured":"Zhao, S., Zhao, L., Zhang, Z., Zhou, E., Metaxas, D.: Global matching with overlapping attention for optical flow estimation. In: Proceedings of the CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01707"},{"key":"2_CR64","unstructured":"Zhao, T., Nevatia, R.: Tracking multiple humans in crowded environment. In: Proceedings of the CVPR, vol.\u00a02 (2004)"},{"key":"2_CR65","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Harley, A.W., Shen, B., Wetzstein, G., Guibas, L.J.: PointOdyssey: a large-scale synthetic dataset for long-term point tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 19855\u201319865 (2023)","DOI":"10.1109\/ICCV51070.2023.01818"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73033-7_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T00:30:30Z","timestamp":1730334630000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73033-7_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,31]]},"ISBN":["9783031730320","9783031730337"],"references-count":65,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73033-7_2","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,31]]},"assertion":[{"value":"31 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}