{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T18:01:06Z","timestamp":1775325666121,"version":"3.50.1"},"publisher-location":"Cham","reference-count":55,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031733369","type":"print"},{"value":"9783031733376","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T00:00:00Z","timestamp":1730332800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T00:00:00Z","timestamp":1730332800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73337-6_10","type":"book-chapter","created":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T23:02:27Z","timestamp":1730329347000},"page":"167-184","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["AFF-ttention! Affordances and\u00a0Attention Models for\u00a0Short-Term Object Interaction Anticipation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-1488-1739","authenticated-orcid":false,"given":"Lorenzo","family":"Mur-Labadia","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6741-844X","authenticated-orcid":false,"given":"Ruben","family":"Martinez-Cantin","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5209-2267","authenticated-orcid":false,"given":"Jose J.","family":"Guerrero","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6034-0432","authenticated-orcid":false,"given":"Giovanni Maria","family":"Farinella","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6911-0302","authenticated-orcid":false,"given":"Antonino","family":"Furnari","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,31]]},"reference":[{"key":"10_CR1","doi-asserted-by":"crossref","unstructured":"Bao, W., et al.: Uncertainty-aware state space transformer for egocentric 3D hand trajectory forecasting. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13702\u201313711 (2023)","DOI":"10.1109\/ICCV51070.2023.01260"},{"key":"10_CR2","unstructured":"Bertasius, G., Wang, H., Torresani, L.: Is space-time attention all you need for video understanding? In: ICML, vol.\u00a02, p.\u00a04 (2021)"},{"key":"10_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"576","DOI":"10.1007\/978-3-030-58571-6_34","volume-title":"Computer Vision \u2013 ECCV 2020","author":"H Bi","year":"2020","unstructured":"Bi, H., Zhang, R., Mao, T., Deng, Z., Wang, Z.: How can I see my future? FvTraj: using first-person view for pedestrian trajectory prediction. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12352, pp. 576\u2013593. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58571-6_34"},{"key":"10_CR4","unstructured":"Chen, G., et\u00a0al.: Internvideo-ego4d: a pack of champion solutions to ego4d challenges. arXiv preprint arXiv:2211.09529 (2022)"},{"key":"10_CR5","doi-asserted-by":"crossref","unstructured":"Chi, H.G., Lee, K., Agarwal, N., Xu, Y., Ramani, K., Choi, C.: Adamsformer for spatial action localization in the future. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 17885\u201317895 (2023)","DOI":"10.1109\/CVPR52729.2023.01715"},{"key":"10_CR6","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"753","DOI":"10.1007\/978-3-030-01225-0_44","volume-title":"Computer Vision \u2013 ECCV 2018","author":"D Damen","year":"2018","unstructured":"Damen, D., et al.: Scaling egocentric vision: the EPIC-KITCHENS dataset. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11208, pp. 753\u2013771. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01225-0_44"},{"key":"10_CR7","unstructured":"Dessalene, E., Devaraj, C., Maynord, M., Fermuller, C., Aloimonos, Y.: Forecasting action through contact representations from first person video. IEEE Trans. Pattern Anal. Mach. Intell. (2021)"},{"key":"10_CR8","doi-asserted-by":"crossref","unstructured":"Do, T.T., Nguyen, A., Reid, I.: Affordancenet: an end-to-end deep learning approach for object affordance detection. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 5882\u20135889. IEEE (2018)","DOI":"10.1109\/ICRA.2018.8460902"},{"key":"10_CR9","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C.: X3D: expanding architectures for efficient video recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 203\u2013213 (2020)","DOI":"10.1109\/CVPR42600.2020.00028"},{"key":"10_CR10","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Fan, H., Malik, J., He, K.: Slowfast networks for video recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6202\u20136211 (2019)","DOI":"10.1109\/ICCV.2019.00630"},{"key":"10_CR11","doi-asserted-by":"publisher","first-page":"401","DOI":"10.1016\/j.jvcir.2017.10.004","volume":"49","author":"A Furnari","year":"2017","unstructured":"Furnari, A., Battiato, S., Grauman, K., Farinella, G.M.: Next-active-object prediction from egocentric videos. J. Vis. Commun. Image Represent. 49, 401\u2013411 (2017)","journal-title":"J. Vis. Commun. Image Represent."},{"issue":"11","key":"10_CR12","doi-asserted-by":"publisher","first-page":"4021","DOI":"10.1109\/TPAMI.2020.2992889","volume":"43","author":"A Furnari","year":"2020","unstructured":"Furnari, A., Farinella, G.M.: Rolling-unrolling LSTMs for action anticipation from first-person video. IEEE Trans. Pattern Anal. Mach. Intell. 43(11), 4021\u20134036 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"2","key":"10_CR13","first-page":"67","volume":"1","author":"JJ Gibson","year":"1977","unstructured":"Gibson, J.J.: The theory of affordances. Hilldale, USA 1(2), 67\u201382 (1977)","journal-title":"Hilldale, USA"},{"key":"10_CR14","doi-asserted-by":"crossref","unstructured":"Girdhar, R., Grauman, K.: Anticipative video transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13505\u201313515 (2021)","DOI":"10.1109\/ICCV48922.2021.01325"},{"key":"10_CR15","doi-asserted-by":"crossref","unstructured":"Girshick, R.: Fast R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1440\u20131448 (2015)","DOI":"10.1109\/ICCV.2015.169"},{"key":"10_CR16","doi-asserted-by":"crossref","unstructured":"Goyal, M., Modi, S., Goyal, R., Gupta, S.: Human hands as probes for interactive object understanding. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3293\u20133303 (2022)","DOI":"10.1109\/CVPR52688.2022.00329"},{"key":"10_CR17","unstructured":"Grauman, K., et\u00a0al.: Ego4d: around the world in 3,000 hours of egocentric video. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18995\u201319012 (2022)"},{"key":"10_CR18","doi-asserted-by":"publisher","first-page":"212","DOI":"10.1016\/j.neucom.2020.12.069","volume":"433","author":"J Jiang","year":"2021","unstructured":"Jiang, J., Nan, Z., Chen, H., Chen, S., Zheng, N.: Predicting short-term next-active-object through visual attention and hand position. Neurocomputing 433, 212\u2013222 (2021)","journal-title":"Neurocomputing"},{"key":"10_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"201","DOI":"10.1007\/978-3-642-33765-9_15","volume-title":"Computer Vision \u2013 ECCV 2012","author":"KM Kitani","year":"2012","unstructured":"Kitani, K.M., Ziebart, B.D., Bagnell, J.A., Hebert, M.: Activity forecasting. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012, Part IV. LNCS, vol. 7575, pp. 201\u2013214. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33765-9_15"},{"issue":"1","key":"10_CR20","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1109\/TPAMI.2015.2430335","volume":"38","author":"HS Koppula","year":"2015","unstructured":"Koppula, H.S., Saxena, A.: Anticipating human activities using object affordances for reactive robotic response. IEEE Trans. Pattern Anal. Mach. Intell. 38(1), 14\u201329 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10_CR21","doi-asserted-by":"crossref","unstructured":"Lee, Y.J., Ghosh, J., Grauman, K.: Discovering important people and objects for egocentric video summarization. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1346\u20131353. IEEE (2012)","DOI":"10.1109\/CVPR.2012.6247820"},{"key":"10_CR22","doi-asserted-by":"crossref","unstructured":"Li, G., Jampani, V., Sun, D., Sevilla-Lara, L.: Locate: localize and transfer object parts for weakly supervised affordance grounding. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10922\u201310931 (2023)","DOI":"10.1109\/CVPR52729.2023.01051"},{"key":"10_CR23","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117\u20132125 (2017)","DOI":"10.1109\/CVPR.2017.106"},{"key":"10_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"704","DOI":"10.1007\/978-3-030-58452-8_41","volume-title":"Computer Vision \u2013 ECCV 2020","author":"M Liu","year":"2020","unstructured":"Liu, M., Tang, S., Li, Y., Rehg, J.M.: Forecasting human-object interaction: joint prediction of motor attention and\u00a0actions in first person video. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020, Part I. LNCS, vol. 12346, pp. 704\u2013721. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_41"},{"key":"10_CR25","doi-asserted-by":"crossref","unstructured":"Liu, S., Tripathi, S., Majumdar, S., Wang, X.: Joint hand motion and interaction hotspots prediction from egocentric videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3282\u20133292 (2022)","DOI":"10.1109\/CVPR52688.2022.00328"},{"key":"10_CR26","doi-asserted-by":"crossref","unstructured":"Luo, H., Zhai, W., Zhang, J., Cao, Y., Tao, D.: Learning visual affordance grounding from demonstration videos. IEEE Trans. Neural Netw. Learn. Syst. (2023)","DOI":"10.1109\/TNNLS.2023.3298638"},{"key":"10_CR27","unstructured":"Marchetti, F., Becattini, F., Seidenari, L., Del\u00a0Bimbo, A.: Multiple trajectory prediction of moving agents with memory augmented networks. IEEE Trans. Pattern Anal. Mach. Intell. (2020)"},{"issue":"1","key":"10_CR28","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1109\/TRO.2007.914848","volume":"24","author":"L Montesano","year":"2008","unstructured":"Montesano, L., Lopes, M., Bernardino, A., Santos-Victor, J.: Learning object affordances: from sensory-motor coordination to imitation. IEEE Trans. Rob. 24(1), 15\u201326 (2008)","journal-title":"IEEE Trans. Rob."},{"key":"10_CR29","doi-asserted-by":"crossref","unstructured":"Mur-Labadia, L., Guerrero, J.J., Martinez-Cantin, R.: Multi-label affordance mapping from egocentric vision. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5238\u20135249 (2023)","DOI":"10.1109\/ICCV51070.2023.00483"},{"key":"10_CR30","doi-asserted-by":"crossref","unstructured":"Mur-Labadia, L., Martinez-Cantin, R., Guerrero, J.J.: Bayesian deep learning for affordance segmentation in images. In: 2023 IEEE International Conference on Robotics and Automation (ICRA). IEEE (2023)","DOI":"10.1109\/ICRA48891.2023.10160606"},{"key":"10_CR31","doi-asserted-by":"crossref","unstructured":"Myers, A., Teo, C.L., Ferm\u00fcller, C., Aloimonos, Y.: Affordance detection of tool parts from geometric features. In: 2015 IEEE International Conference on Robotics and Automation (ICRA), pp. 1374\u20131381. IEEE (2015)","DOI":"10.1109\/ICRA.2015.7139369"},{"key":"10_CR32","doi-asserted-by":"crossref","unstructured":"Nagarajan, T., Feichtenhofer, C., Grauman, K.: Grounded human-object interaction hotspots from video. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8688\u20138697 (2019)","DOI":"10.1109\/ICCV.2019.00878"},{"key":"10_CR33","doi-asserted-by":"crossref","unstructured":"Nagarajan, T., Li, Y., Feichtenhofer, C., Grauman, K.: Ego-topo: environment affordances from egocentric video. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 163\u2013172 (2020)","DOI":"10.1109\/CVPR42600.2020.00024"},{"key":"10_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"558","DOI":"10.1007\/978-3-031-19830-4_32","volume-title":"Computer Vision - ECCV 2022","author":"M Nawhal","year":"2022","unstructured":"Nawhal, M., Jyothi, A.A., Mori, G.: Rethinking learning approaches for long-term action anticipation. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13694, pp. 558\u2013576. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19830-4_32"},{"key":"10_CR35","doi-asserted-by":"crossref","unstructured":"Nguyen, A., Kanoulas, D., Caldwell, D.G., Tsagarakis, N.G.: Object-based affordances detection with convolutional neural networks and dense conditional random fields. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5908\u20135915. IEEE (2017)","DOI":"10.1109\/IROS.2017.8206484"},{"key":"10_CR36","unstructured":"Oquab, M., et\u00a0al.: Dinov2: learning robust visual features without supervision. Trans. Mach. Learn. Res. (2024)"},{"key":"10_CR37","doi-asserted-by":"crossref","unstructured":"Park, H.S., Hwang, J.J., Niu, Y., Shi, J.: Egocentric future localization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4697\u20134705 (2016)","DOI":"10.1109\/CVPR.2016.508"},{"key":"10_CR38","doi-asserted-by":"crossref","unstructured":"Pasca, R.G., Gavryushin, A., Kuo, Y.L., Hilliges, O., Wang, X.: Summarize the past to predict the future: natural language descriptions of context boost multimodal object interaction. arXiv preprint arXiv:2301.09209 (2023)","DOI":"10.1109\/CVPR52733.2024.01731"},{"key":"10_CR39","unstructured":"Plizzari, C., et al.: An outlook into the future of egocentric vision. arXiv preprint arXiv:2308.07123 (2023)"},{"key":"10_CR40","doi-asserted-by":"crossref","unstructured":"Plizzari, C., Perrett, T., Caputo, B., Damen, D.: What can a cook in Italy teach a mechanic in India? Action recognition generalisation over scenarios and locations. In: ICCV 2023 (2023)","DOI":"10.1109\/ICCV51070.2023.01256"},{"key":"10_CR41","doi-asserted-by":"crossref","unstructured":"Pramanick, S., et al.: Egovlpv2: egocentric video-language pre-training with fusion in the backbone. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5285\u20135297 (2023)","DOI":"10.1109\/ICCV51070.2023.00487"},{"key":"10_CR42","doi-asserted-by":"crossref","unstructured":"Ragusa, F., Farinella, G.M., Furnari, A.: Stillfast: an end-to-end approach for short-term object interaction anticipation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3635\u20133644 (2023)","DOI":"10.1109\/CVPRW59228.2023.00371"},{"key":"10_CR43","doi-asserted-by":"crossref","unstructured":"Ragusa, F., Furnari, A., Livatino, S., Farinella, G.M.: The meccano dataset: understanding human-object interactions from egocentric videos in an industrial-like domain. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1569\u20131578 (2021)","DOI":"10.1109\/WACV48630.2021.00161"},{"key":"10_CR44","doi-asserted-by":"crossref","unstructured":"Rhinehart, N., Kitani, K.M.: Learning action maps of large environments via first-person vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580\u2013588 (2016)","DOI":"10.1109\/CVPR.2016.69"},{"key":"10_CR45","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2021.103252","volume":"211","author":"I Rodin","year":"2021","unstructured":"Rodin, I., Furnari, A., Mavroeidis, D., Farinella, G.M.: Predicting the future from first person (egocentric) vision: a survey. Comput. Vis. Image Underst. 211, 103252 (2021)","journal-title":"Comput. Vis. Image Underst."},{"key":"10_CR46","doi-asserted-by":"crossref","unstructured":"Roy, D., Rajendiran, R., Fernando, B.: Interaction region visual transformer for egocentric action anticipation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 6740\u20136750 (2024)","DOI":"10.1109\/WACV57701.2024.00660"},{"key":"10_CR47","doi-asserted-by":"crossref","unstructured":"Shan, D., Geng, J., Shu, M., Fouhey, D.F.: Understanding human hands in contact at internet scale. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9869\u20139878 (2020)","DOI":"10.1109\/CVPR42600.2020.00989"},{"key":"10_CR48","unstructured":"Team, E.: Short-Term object-interaction Anticipation quickstart (2023). https:\/\/colab.research.google.com\/drive\/1Ok_6F1O6K8kX1S4sEnU62HoOBw_CPngR?usp=sharing. Accessed 03 Mar 2024"},{"key":"10_CR49","doi-asserted-by":"crossref","unstructured":"Thakur, S., Beyan, C., Morerio, P., Murino, V., Del\u00a0Bue, A.: Enhancing next active object-based egocentric action anticipation with guided attention. In: International Conference on Image Processing (2023)","DOI":"10.2139\/ssrn.4635951"},{"key":"10_CR50","unstructured":"Thakur, S., Beyan, C., Morerio, P., Murino, V., Del\u00a0Bue, A.: Guided attention for next active object@ ego4d STA challenge. In: CVPR23 EGO4D Workshop STA Challenge (2023)"},{"key":"10_CR51","doi-asserted-by":"crossref","unstructured":"Thakur, S., Beyan, C., Morerio, P., Murino, V., Del\u00a0Bue, A.: Leveraging next-active objects for context-aware anticipation in egocentric videos. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 8657\u20138666 (2024)","DOI":"10.1109\/WACV57701.2024.00846"},{"key":"10_CR52","unstructured":"Tong, Z., Song, Y., Wang, J., Wang, L.: VideoMAE: masked autoencoders are data-efficient learners for self-supervised video pre-training. In: Advances in Neural Information Processing Systems, vol. 35, pp. 10078\u201310093 (2022)"},{"key":"10_CR53","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"10_CR54","doi-asserted-by":"crossref","unstructured":"Zatsarynna, O., Abu\u00a0Farha, Y., Gall, J.: Multi-modal temporal convolutional network for anticipating actions in egocentric videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2249\u20132258 (2021)","DOI":"10.1109\/CVPRW53098.2021.00254"},{"key":"10_CR55","doi-asserted-by":"crossref","unstructured":"Zhong, Z., Schneider, D., Voit, M., Stiefelhagen, R., Beyerer, J.: Anticipative feature fusion transformer for multi-modal action anticipation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 6068\u20136077 (2023)","DOI":"10.1109\/WACV56688.2023.00601"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73337-6_10","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T23:04:58Z","timestamp":1730329498000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73337-6_10"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,31]]},"ISBN":["9783031733369","9783031733376"],"references-count":55,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73337-6_10","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,31]]},"assertion":[{"value":"31 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}