{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,28]],"date-time":"2025-05-28T04:16:54Z","timestamp":1748405814130,"version":"3.41.0"},"publisher-location":"Cham","reference-count":60,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031924590","type":"print"},{"value":"9783031924606","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-92460-6_1","type":"book-chapter","created":{"date-parts":[[2025,5,27]],"date-time":"2025-05-27T16:36:45Z","timestamp":1748363805000},"page":"1-18","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["CEIA: CLIP-Based Event-Image Alignment for\u00a0Open-World Event-Based Understanding"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-9003-9645","authenticated-orcid":false,"given":"Wenhao","family":"Xu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1042-8903","authenticated-orcid":false,"given":"Wenming","family":"Weng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0788-8826","authenticated-orcid":false,"given":"Yueyi","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9787-7460","authenticated-orcid":false,"given":"Zhiwei","family":"Xiong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,12]]},"reference":[{"key":"1_CR1","doi-asserted-by":"crossref","unstructured":"Alzugaray, I., Chli, M.: ACE: an efficient asynchronous corner tracker for event cameras. In: 2018 International Conference on 3D Vision (3DV), pp. 653\u2013661. IEEE (2018)","DOI":"10.1109\/3DV.2018.00080"},{"key":"1_CR2","unstructured":"Berner, R., Brandli, C., Yang, M., Liu, S.C., Delbruck, T.: A $$240\\times 180$$ 10 mw 12 us latency sparse-output vision sensor for mobile applications. In: 2013 Symposium on VLSI Circuits, pp. C186\u2013C187. IEEE (2013)"},{"key":"1_CR3","doi-asserted-by":"publisher","first-page":"9084","DOI":"10.1109\/TIP.2020.3023597","volume":"29","author":"Y Bi","year":"2020","unstructured":"Bi, Y., Chadha, A., Abbas, A., Bourtsoulatze, E., Andreopoulos, Y.: Graph-based spatio-temporal feature learning for neuromorphic vision sensing. IEEE Trans. Image Process. 29, 9084\u20139098 (2020)","journal-title":"IEEE Trans. Image Process."},{"issue":"10","key":"1_CR4","doi-asserted-by":"publisher","first-page":"2333","DOI":"10.1109\/JSSC.2014.2342715","volume":"49","author":"C Brandli","year":"2014","unstructured":"Brandli, C., Berner, R., Yang, M., Liu, S.C., Delbruck, T.: A $$240\\times 180$$ 130 db 3 $$\\upmu $$s latency global shutter spatiotemporal vision sensor. IEEE J. Solid-State Circuits 49(10), 2333\u20132341 (2014)","journal-title":"IEEE J. Solid-State Circuits"},{"key":"1_CR5","doi-asserted-by":"crossref","unstructured":"Brooks, T., Holynski, A., Efros, A.A.: Instructpix2pix: learning to follow image editing instructions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18392\u201318402 (2023)","DOI":"10.1109\/CVPR52729.2023.01764"},{"key":"1_CR6","doi-asserted-by":"crossref","unstructured":"Das\u00a0Biswas, S., Kosta, A., Liyanagedera, C., Apolinario, M., Roy, K.: Halsie: hybrid approach to learning segmentation by simultaneously exploiting image and event modalities. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 5964\u20135974 (2024)","DOI":"10.1109\/WACV57701.2024.00586"},{"key":"1_CR7","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. IEEE (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"1_CR8","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth $$16 \\times 16$$ words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"1_CR9","doi-asserted-by":"crossref","unstructured":"Farahani, A., Voghoei, S., Rasheed, K., Arabnia, H.R.: A brief review of domain adaptation. In: Advances in Data Science and Information Engineering: Proceedings from ICDATA 2020 and IKE 2020, pp. 877\u2013894 (2021)","DOI":"10.1007\/978-3-030-71704-9_65"},{"key":"1_CR10","unstructured":"Fei-Fei, L., Fergus, R., Perona, P.: Learning generative visual models from few training examples: an incremental Bayesian approach tested on 101 object categories. In: 2004 Conference on Computer Vision and Pattern Recognition Workshop, p. 178. IEEE (2004)"},{"issue":"1","key":"1_CR11","doi-asserted-by":"publisher","first-page":"154","DOI":"10.1109\/TPAMI.2020.3008413","volume":"44","author":"G Gallego","year":"2020","unstructured":"Gallego, G., et al.: Event-based vision: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 154\u2013180 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1_CR12","doi-asserted-by":"crossref","unstructured":"Gehrig, D., Loquercio, A., Derpanis, K.G., Scaramuzza, D.: End-to-end learning of representations for asynchronous event-based data. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5633\u20135643 (2019)","DOI":"10.1109\/ICCV.2019.00573"},{"key":"1_CR13","doi-asserted-by":"crossref","unstructured":"Glover, A., Vasco, V., Bartolozzi, C.: A controlled-delay event camera framework for on-line robotics. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 2178\u20132183. IEEE (2018)","DOI":"10.1109\/ICRA.2018.8460541"},{"issue":"12","key":"1_CR14","doi-asserted-by":"publisher","first-page":"4338","DOI":"10.1109\/TPAMI.2020.3005434","volume":"43","author":"Y Guo","year":"2020","unstructured":"Guo, Y., Wang, H., Hu, Q., Liu, H., Liu, L., Bennamoun, M.: Deep learning for 3D point clouds: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 43(12), 4338\u20134364 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1_CR15","doi-asserted-by":"crossref","unstructured":"Guzhov, A., Raue, F., Hees, J., Dengel, A.: Audioclip: extending clip to image, text and audio. In: ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 976\u2013980. IEEE (2022)","DOI":"10.1109\/ICASSP43922.2022.9747631"},{"key":"1_CR16","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"1_CR17","doi-asserted-by":"crossref","unstructured":"Hegde, D., Valanarasu, J.M.J., Patel, V.: Clip goes 3D: leveraging prompt tuning for language grounded 3D recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2028\u20132038 (2023)","DOI":"10.1109\/ICCVW60793.2023.00217"},{"key":"1_CR18","unstructured":"Hu, E.J., et al.: Lora: low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)"},{"issue":"22","key":"1_CR19","doi-asserted-by":"publisher","first-page":"35864","DOI":"10.1364\/OE.437944","volume":"29","author":"X Huang","year":"2021","unstructured":"Huang, X., Zhang, Y., Xiong, Z.: High-speed structured light based 3D scanning using an event camera. Opt. Express 29(22), 35864\u201335876 (2021)","journal-title":"Opt. Express"},{"key":"1_CR20","doi-asserted-by":"crossref","unstructured":"Huang, X., Zhang, Y., Xiong, Z.: Progressive spatio-temporal alignment for efficient event-based motion estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1537\u20131546 (2023)","DOI":"10.1109\/CVPR52729.2023.00154"},{"key":"1_CR21","unstructured":"Jia, C., et al.: Scaling up visual and vision-language representation learning with noisy text supervision. In: International Conference on Machine Learning, pp. 4904\u20134916. PMLR (2021)"},{"key":"1_CR22","doi-asserted-by":"crossref","unstructured":"Kim, J., Bae, J., Park, G., Zhang, D., Kim, Y.M.: N-imagenet: towards robust, fine-grained object recognition with event cameras. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2146\u20132156 (2021)","DOI":"10.1109\/ICCV48922.2021.00215"},{"issue":"4","key":"1_CR23","doi-asserted-by":"publisher","DOI":"10.1088\/2634-4386\/ac9b86","volume":"2","author":"Y Kim","year":"2022","unstructured":"Kim, Y., Chough, J., Panda, P.: Beyond classification: directly training spiking neural networks for semantic segmentation. Neuromorphic Comput. Eng. 2(4), 044015 (2022)","journal-title":"Neuromorphic Comput. Eng."},{"key":"1_CR24","doi-asserted-by":"crossref","unstructured":"Klenk, S., Bonello, D., Koestler, L., Araslanov, N., Cremers, D.: Masked event modeling: self-supervised pretraining for event cameras. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 2378\u20132388 (2024)","DOI":"10.1109\/WACV57701.2024.00237"},{"key":"1_CR25","unstructured":"Krizhevsky, A.: Learning multiple layers of features from tiny images. Master\u2019s thesis, University of Tront (2009)"},{"issue":"7","key":"1_CR26","doi-asserted-by":"publisher","first-page":"1346","DOI":"10.1109\/TPAMI.2016.2574707","volume":"39","author":"X Lagorce","year":"2016","unstructured":"Lagorce, X., Orchard, G., Galluppi, F., Shi, B.E., Benosman, R.B.: Hots: a hierarchy of event-based time-surfaces for pattern recognition. IEEE Trans. Pattern Anal. Mach. Intell. 39(7), 1346\u20131359 (2016)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1_CR27","doi-asserted-by":"crossref","unstructured":"Lee, K.H., Chen, X., Hua, G., Hu, H., He, X.: Stacked cross attention for image-text matching. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 201\u2013216 (2018)","DOI":"10.1007\/978-3-030-01225-0_13"},{"key":"1_CR28","doi-asserted-by":"publisher","first-page":"309","DOI":"10.3389\/fnins.2017.00309","volume":"11","author":"H Li","year":"2017","unstructured":"Li, H., Liu, H., Ji, X., Li, G., Shi, L.: Cifar10-DVS: an event-stream dataset for object classification. Front. Neurosci. 11, 309 (2017)","journal-title":"Front. Neurosci."},{"key":"1_CR29","unstructured":"Li, J., Li, D., Savarese, S., Hoi, S.: Blip-2: bootstrapping language-image pre-training with frozen image encoders and large language models. arXiv preprint arXiv:2301.12597 (2023)"},{"key":"1_CR30","doi-asserted-by":"crossref","unstructured":"Li, N., Bhat, A., Raychowdhury, A.: E-track: eye tracking with event camera for extended reality (XR) applications. In: 2023 IEEE 5th International Conference on Artificial Intelligence Circuits and Systems (AICAS), pp.\u00a01\u20135. IEEE (2023)","DOI":"10.1109\/AICAS57966.2023.10168551"},{"key":"1_CR31","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"695","DOI":"10.1007\/978-3-030-58598-3_41","volume-title":"Computer Vision \u2013 ECCV 2020","author":"S Lin","year":"2020","unstructured":"Lin, S., et al.: Learning event-driven video deblurring and interpolation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12353, pp. 695\u2013710. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58598-3_41"},{"key":"1_CR32","unstructured":"Loshchilov, I., Hutter, F.: SGDR: stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983 (2016)"},{"key":"1_CR33","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"issue":"2","key":"1_CR34","doi-asserted-by":"publisher","first-page":"3515","DOI":"10.1109\/LRA.2022.3145053","volume":"7","author":"N Messikommer","year":"2022","unstructured":"Messikommer, N., Gehrig, D., Gehrig, M., Scaramuzza, D.: Bridging the gap between events and frames through unsupervised domain adaptation. IEEE Robot. Autom. Lett. 7(2), 3515\u20133522 (2022)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"1_CR35","doi-asserted-by":"crossref","unstructured":"Ni, B., et al.: Expanding language-image pretrained models for general video recognition. In: European Conference on Computer Vision, pp. 1\u201318. Springer, Cham (2022)","DOI":"10.1007\/978-3-031-19772-7_1"},{"key":"1_CR36","unstructured":"van den Oord, A., Li, Y., Vinyals, O.: Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018)"},{"key":"1_CR37","doi-asserted-by":"publisher","first-page":"437","DOI":"10.3389\/fnins.2015.00437","volume":"9","author":"G Orchard","year":"2015","unstructured":"Orchard, G., Jayawant, A., Cohen, G.K., Thakor, N.: Converting static image datasets to spiking neuromorphic datasets using saccades. Front. Neurosci. 9, 437 (2015)","journal-title":"Front. Neurosci."},{"key":"1_CR38","doi-asserted-by":"crossref","unstructured":"Paikin, G., Ater, Y., Shaul, R., Soloveichik, E.: Efi-net: video frame interpolation from fusion of events and frames. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1291\u20131301 (2021)","DOI":"10.1109\/CVPRW53098.2021.00142"},{"key":"1_CR39","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"1_CR40","doi-asserted-by":"crossref","unstructured":"Rao, Y., et al.: Denseclip: language-guided dense prediction with context-aware prompting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18082\u201318091 (2022)","DOI":"10.1109\/CVPR52688.2022.01755"},{"key":"1_CR41","doi-asserted-by":"crossref","unstructured":"Rebecq, H., Ranftl, R., Koltun, V., Scaramuzza, D.: Events-to-video: bringing modern computer vision to event cameras. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3857\u20133866 (2019)","DOI":"10.1109\/CVPR.2019.00398"},{"key":"1_CR42","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"1_CR43","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"1_CR44","doi-asserted-by":"crossref","unstructured":"Sun, Z., Messikommer, N., Gehrig, D., Scaramuzza, D.: ESS: learning event-based semantic segmentation from still images. In: European Conference on Computer Vision, pp. 341\u2013357. Springer, Cham (2022)","DOI":"10.1007\/978-3-031-19830-4_20"},{"key":"1_CR45","doi-asserted-by":"crossref","unstructured":"Tulyakov, S., Bochicchio, A., Gehrig, D., Georgoulis, S., Li, Y., Scaramuzza, D.: Time lens++: event-based frame interpolation with parametric non-linear flow and multi-scale fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 17755\u201317764 (2022)","DOI":"10.1109\/CVPR52688.2022.01723"},{"key":"1_CR46","doi-asserted-by":"crossref","unstructured":"Wang, J., Weng, W., Zhang, Y., Xiong, Z.: Unsupervised video deraining with an event camera. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10831\u201310840 (2023)","DOI":"10.1109\/ICCV51070.2023.00994"},{"key":"1_CR47","doi-asserted-by":"crossref","unstructured":"Weng, W., Zhang, Y., Xiong, Z.: Event-based video reconstruction using transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2563\u20132572 (2021)","DOI":"10.1109\/ICCV48922.2021.00256"},{"key":"1_CR48","doi-asserted-by":"crossref","unstructured":"Weng, W., Zhang, Y., Xiong, Z.: Event-based blurry frame interpolation under blind exposure. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1588\u20131598 (2023)","DOI":"10.1109\/CVPR52729.2023.00159"},{"key":"1_CR49","unstructured":"Wu, Z., Liu, X., Gilitschenski, I.: Eventclip: adapting clip for event-based object recognition. arXiv preprint arXiv:2306.06354 (2023)"},{"key":"1_CR50","doi-asserted-by":"publisher","first-page":"1145","DOI":"10.1109\/TCI.2022.3228747","volume":"8","author":"Z Xiao","year":"2022","unstructured":"Xiao, Z., Weng, W., Zhang, Y., Xiong, Z.: Eva $$^2$$: event-assisted video frame interpolation via cross-modal alignment and aggregation. IEEE Trans. Comput. Imaging 8, 1145\u20131158 (2022)","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"1_CR51","doi-asserted-by":"crossref","unstructured":"Xue, L., et al.: Ulip: learning a unified representation of language, images, and point clouds for 3D understanding. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1179\u20131189 (2023)","DOI":"10.1109\/CVPR52729.2023.00120"},{"key":"1_CR52","doi-asserted-by":"crossref","unstructured":"Yang, Y., Pan, L., Liu, L.: Event camera data pre-training. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10699\u201310709 (2023)","DOI":"10.1109\/ICCV51070.2023.00982"},{"key":"1_CR53","doi-asserted-by":"crossref","unstructured":"Yu, Z., et al.: Training weakly supervised video frame interpolation with events. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14589\u201314598 (2021)","DOI":"10.1109\/ICCV48922.2021.01432"},{"key":"1_CR54","unstructured":"Yuan, L., et\u00a0al.: Florence: a new foundation model for computer vision. arXiv preprint arXiv:2111.11432 (2021)"},{"key":"1_CR55","doi-asserted-by":"crossref","unstructured":"Zhang, R., et al.: Pointclip: point cloud understanding by clip. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8552\u20138562 (2022)","DOI":"10.1109\/CVPR52688.2022.00836"},{"key":"1_CR56","unstructured":"Zhang, S., et\u00a0al.: OPT: open pre-trained transformer language models. arXiv preprint arXiv:2205.01068 (2022)"},{"key":"1_CR57","doi-asserted-by":"crossref","unstructured":"Zhang, X., Yu, L.: Unifying motion deblurring and frame interpolation with events. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 17765\u201317774 (2022)","DOI":"10.1109\/CVPR52688.2022.01724"},{"issue":"5","key":"1_CR58","doi-asserted-by":"publisher","first-page":"1284","DOI":"10.1007\/s11263-023-01754-5","volume":"131","author":"C Zhou","year":"2023","unstructured":"Zhou, C., et al.: Deblurring low-light images with events. Int. J. Comput. Vis. 131(5), 1284\u20131298 (2023)","journal-title":"Int. J. Comput. Vis."},{"key":"1_CR59","unstructured":"Zhou, J., Zheng, X., Lyu, Y., Wang, L.: E-clip: towards label-efficient event-based open-world understanding by clip. arXiv preprint arXiv:2308.03135 (2023)"},{"key":"1_CR60","doi-asserted-by":"crossref","unstructured":"Zhu, A.Z., Yuan, L., Chaney, K., Daniilidis, K.: Unsupervised event-based learning of optical flow, depth, and egomotion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 989\u2013997 (2019)","DOI":"10.1109\/CVPR.2019.00108"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-92460-6_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,27]],"date-time":"2025-05-27T16:37:08Z","timestamp":1748363828000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-92460-6_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031924590","9783031924606"],"references-count":60,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-92460-6_1","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"12 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}