{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,14]],"date-time":"2026-02-14T10:34:45Z","timestamp":1771065285219,"version":"3.50.1"},"reference-count":71,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2025,7,22]],"date-time":"2025-07-22T00:00:00Z","timestamp":1753142400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,7,22]],"date-time":"2025-07-22T00:00:00Z","timestamp":1753142400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key R&D Program of China","doi-asserted-by":"crossref","award":["2021YFF0900500"],"award-info":[{"award-number":["2021YFF0900500"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62441202"],"award-info":[{"award-number":["62441202"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["U22B2035"],"award-info":[{"award-number":["U22B2035"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62027804"],"award-info":[{"award-number":["62027804"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62088102"],"award-info":[{"award-number":["62088102"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s11263-025-02488-2","type":"journal-article","created":{"date-parts":[[2025,7,22]],"date-time":"2025-07-22T16:06:21Z","timestamp":1753200381000},"page":"7332-7351","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["High-Rate Monocular Depth Estimation via Cross Frame-Rate Collaboration of Frames and Events"],"prefix":"10.1007","volume":"133","author":[{"given":"Xu","family":"Liu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9660-3636","authenticated-orcid":false,"given":"Xiaopeng","family":"Fan","sequence":"additional","affiliation":[]},{"given":"Jianing","family":"Li","sequence":"additional","affiliation":[]},{"given":"Dianze","family":"Li","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Zhengyu","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Yonghong","family":"Tian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,22]]},"reference":[{"key":"2488_CR1","first-page":"20014","volume":"34","author":"A Ali","year":"2021","unstructured":"Ali, A., Touvron, H., Caron, M., Bojanowski, P., Douze, M., Joulin, A., Laptev, I., Neverova, N., Synnaeve, G., Verbeek, J., et al. (2021). Xcit: Cross-covariance image transformers. Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), 34, 20014\u201320027.","journal-title":"Proceedings of the Advances in Neural Information Processing Systems (NeurIPS)"},{"key":"2488_CR2","unstructured":"Ballas, N., Yao, L., Pal, C., & Courville, A. (2015) Delving deeper into convolutional networks for learning video representations. arXiv preprint arXiv:1511.06432"},{"key":"2488_CR3","doi-asserted-by":"crossref","unstructured":"Bodden, L., Schwaiger, F., Ha, D. B., Kreuzberg, L., & Behnke, S. (2024) Spiking centernet: A distillation-boosted spiking neural network for object detection. arXiv preprint arXiv:2402.01287","DOI":"10.1109\/IJCNN60899.2024.10650418"},{"issue":"10","key":"2488_CR4","doi-asserted-by":"publisher","first-page":"2333","DOI":"10.1109\/JSSC.2014.2342715","volume":"49","author":"C Brandli","year":"2014","unstructured":"Brandli, C., Berner, R., Yang, M., Liu, S. C., & Delbruck, T. (2014). A 240$$\\times $$ 180 130 db 3 $$\\mu $$s latency global shutter spatiotemporal vision sensor. IEEE journal of solid-state circuits (JSSC), 49(10), 2333\u20132341.","journal-title":"IEEE journal of solid-state circuits (JSSC)"},{"key":"2488_CR5","doi-asserted-by":"crossref","unstructured":"Chen, N. F. (2018) Pseudo-labels for supervised learning on dynamic vision sensor data, applied to object detection under ego-motion. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), (pp 644\u2013653)","DOI":"10.1109\/CVPRW.2018.00107"},{"key":"2488_CR6","doi-asserted-by":"crossref","unstructured":"Cho, H., & Yoon, K. J. (2022) Selection and cross similarity for event-image deep stereo. In: Proceedings of the European Conference on Computer Vision (ECCV), Springer, (pp 470\u2013486)","DOI":"10.1007\/978-3-031-19824-3_28"},{"key":"2488_CR7","first-page":"1","volume":"71","author":"M Cui","year":"2022","unstructured":"Cui, M., Zhu, Y., Liu, Y., Liu, Y., Chen, G., & Huang, K. (2022). Dense depth-map estimation based on fusion of event camera and sparse lidar. IEEE Transactions on Instrumentation and Measurement (TIM), 71, 1\u201311.","journal-title":"IEEE Transactions on Instrumentation and Measurement (TIM)"},{"key":"2488_CR8","doi-asserted-by":"crossref","unstructured":"Dai, J., Qi, H., Xiong, Y., Li, Y., Zhang, G., Hu, H., & Wei, Y. (2017) Deformable convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), (pp 764\u2013773)","DOI":"10.1109\/ICCV.2017.89"},{"key":"2488_CR9","doi-asserted-by":"crossref","unstructured":"Devulapally, A., Khan, M. F. F., Advani, S., & Narayanan, V. (2024) Multi-modal fusion of event and rgb for monocular depth estimation using a unified transformer-based architecture. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), (pp 2081\u20132089)","DOI":"10.1109\/CVPRW63382.2024.00213"},{"issue":"11","key":"2488_CR10","first-page":"8261","volume":"44","author":"P Duan","year":"2021","unstructured":"Duan, P., Wang, Z. W., Shi, B., Cossairt, O., Huang, T., & Katsaggelos, A. K. (2021). Guided event filtering: Synergy between intensity images and neuromorphic events for high performance imaging. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 44(11), 8261\u20138275.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)"},{"key":"2488_CR11","unstructured":"Eigen, D., Puhrsch, C., & Fergus, R. (2014) Depth map prediction from a single image using a multi-scale deep network. Proceedings of the Advances in Neural Information Processing Systems (NeurIPS) 27"},{"key":"2488_CR12","doi-asserted-by":"crossref","unstructured":"Fan, Y., Zhang, W., Liu, C., Li, M., & Lu, W. (2024) Sfod: Spiking fusion object detector. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 17191\u201317200)","DOI":"10.1109\/CVPR52733.2024.01627"},{"key":"2488_CR13","doi-asserted-by":"crossref","unstructured":"Gallego, .G, Rebecq, H., & Scaramuzza, D. (2018) A unifying contrast maximization framework for event cameras, with applications to motion, depth, and optical flow estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 3867\u20133876)","DOI":"10.1109\/CVPR.2018.00407"},{"issue":"1","key":"2488_CR14","doi-asserted-by":"publisher","first-page":"154","DOI":"10.1109\/TPAMI.2020.3008413","volume":"44","author":"G Gallego","year":"2020","unstructured":"Gallego, G., Delbr\u00fcck, T., Orchard, G., Bartolozzi, C., Taba, B., Censi, A., Leutenegger, S., Davison, A. J., Conradt, J., Daniilidis, K., et al. (2020). Event-based vision: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 44(1), 154\u2013180.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)"},{"issue":"3","key":"2488_CR15","doi-asserted-by":"publisher","first-page":"8217","DOI":"10.1109\/LRA.2022.3186770","volume":"7","author":"L Gao","year":"2022","unstructured":"Gao, L., Liang, Y., Yang, J., Wu, S., Wang, C., Chen, J., & Kneip, L. (2022). Vector: A versatile event-centric benchmark for multi-sensor slam. IEEE Robotics and Automation Letters (RAL), 7(3), 8217\u20138224.","journal-title":"IEEE Robotics and Automation Letters (RAL)"},{"key":"2488_CR16","doi-asserted-by":"crossref","unstructured":"Garg, R., Bg, V. K., Carneiro, G., & Reid, I. (2016) Unsupervised cnn for single view depth estimation: Geometry to the rescue. In: Proceedings of the European Conference on Computer Vision (ECCV), Springer, (pp 740\u2013756)","DOI":"10.1007\/978-3-319-46484-8_45"},{"issue":"2","key":"2488_CR17","doi-asserted-by":"publisher","first-page":"2822","DOI":"10.1109\/LRA.2021.3060707","volume":"6","author":"D Gehrig","year":"2021","unstructured":"Gehrig, D., R\u00fcegg, M., Gehrig, M., Hidalgo-Carri\u00f3, J., & Scaramuzza, D. (2021). Combining events and frames using recurrent asynchronous multimodal networks for monocular depth prediction. IEEE Robotics and Automation Letters (RAL), 6(2), 2822\u20132829.","journal-title":"IEEE Robotics and Automation Letters (RAL)"},{"issue":"3","key":"2488_CR18","doi-asserted-by":"publisher","first-page":"4947","DOI":"10.1109\/LRA.2021.3068942","volume":"6","author":"M Gehrig","year":"2021","unstructured":"Gehrig, M., Aarents, W., Gehrig, D., & Scaramuzza, D. (2021). Dsec: A stereo event camera dataset for driving scenarios. IEEE Robotics and Automation Letters (RAL), 6(3), 4947\u20134954.","journal-title":"IEEE Robotics and Automation Letters (RAL)"},{"key":"2488_CR19","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac\u00a0Aodha, O., & Brostow, G. J. (2017) Unsupervised monocular depth estimation with left-right consistency. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 270\u2013279)","DOI":"10.1109\/CVPR.2017.699"},{"key":"2488_CR20","doi-asserted-by":"crossref","unstructured":"Hamaguchi, R., Furukawa, Y., Onishi, M., & Sakurada, K. (2023) Hierarchical neural memory network for low latency event processing. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 22867\u201322876)","DOI":"10.1109\/CVPR52729.2023.02190"},{"key":"2488_CR21","unstructured":"Hendrycks, D., & Gimpel, K. (2016) Bridging nonlinearities and stochastic regularizers with gaussian error linear units. CoRR abs\/1606.08415"},{"key":"2488_CR22","doi-asserted-by":"crossref","unstructured":"Hidalgo-Carri\u00f3, J., Gehrig, D., & Scaramuzza, D. (2020) Learning monocular dense depth from events. In: Proceedings of the IEEE International Conference on 3D Vision (3DV), IEEE, (pp 534\u2013542)","DOI":"10.1109\/3DV50981.2020.00063"},{"key":"2488_CR23","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., & Sun, G. (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 7132\u20137141)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"2488_CR24","doi-asserted-by":"crossref","unstructured":"Huang, X., & Belongie, S. (2017) Arbitrary style transfer in real-time with adaptive instance normalization. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), (pp 1501\u20131510)","DOI":"10.1109\/ICCV.2017.167"},{"key":"2488_CR25","doi-asserted-by":"crossref","unstructured":"Jiang, Z., Xia, P., Huang, K., Stechele, W., Chen, G., Bing, Z., & Knoll, A. (2019) Mixed frame-\/event-driven fast pedestrian detection. In: Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), IEEE, (pp 8332\u20138338)","DOI":"10.1109\/ICRA.2019.8793924"},{"key":"2488_CR26","doi-asserted-by":"crossref","unstructured":"Kim, H., Leutenegger, S., & Davison, A. J. (2016) Real-time 3d reconstruction and 6-dof tracking with an event camera. In: Proceedings of the European Conference on Computer Vision (ECCV), Springer, (pp 349\u2013364)","DOI":"10.1007\/978-3-319-46466-4_21"},{"key":"2488_CR27","unstructured":"Kingma, D. P., & Ba, J. (2014) Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980"},{"key":"2488_CR28","doi-asserted-by":"crossref","unstructured":"Klenk, S., Chui, J., Demmel N., & Cremers, D. (2021) Tum-vie: The tum stereo visual-inertial event dataset. In: Proceedings of the IEEE International Conference on Intelligent Robots and Systems (IROS), eprint2108.07329","DOI":"10.1109\/IROS51168.2021.9636728"},{"issue":"9","key":"2488_CR29","doi-asserted-by":"publisher","first-page":"2265","DOI":"10.1007\/s11263-022-01641-5","volume":"130","author":"S Lee","year":"2022","unstructured":"Lee, S., Rameau, F., Im, S., & Kweon, I. S. (2022). Self-supervised monocular depth and motion learning in dynamic scenes: Semantic prior to rescue. International Journal of Computer Vision (IJCV), 130(9), 2265\u20132285.","journal-title":"International Journal of Computer Vision (IJCV)"},{"issue":"11","key":"2488_CR30","doi-asserted-by":"publisher","first-page":"14020","DOI":"10.1109\/TPAMI.2023.3298925","volume":"45","author":"D Li","year":"2023","unstructured":"Li, D., Li, J., & Tian, Y. (2023). Sodformer: Streaming object detection with transformer using events and frames. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 45(11), 14020\u201314037.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)"},{"key":"2488_CR31","doi-asserted-by":"crossref","unstructured":"Li, J., Dong, S., Yu, Z., Tian, Y., & Huang, T. (2019) Event-based vision enhanced: A joint detection framework in autonomous driving. In: Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), (pp 1396\u20131401)","DOI":"10.1109\/ICME.2019.00242"},{"issue":"4","key":"2488_CR32","doi-asserted-by":"publisher","first-page":"1742","DOI":"10.1109\/TNNLS.2021.3061122","volume":"34","author":"J Li","year":"2021","unstructured":"Li, J., Fu, Y., Dong, S., Yu, Z., Huang, T., & Tian, Y. (2021). Asynchronous spatiotemporal spike metric for event cameras. IEEE Transactions on Neural Networks and Learning Systems (TNNLS), 34(4), 1742\u20131753.","journal-title":"IEEE Transactions on Neural Networks and Learning Systems (TNNLS)"},{"key":"2488_CR33","doi-asserted-by":"crossref","unstructured":"Li, Z., & Snavely, N. (2018) Megadepth: Learning single-view depth prediction from internet photos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 2041\u20132050)","DOI":"10.1109\/CVPR.2018.00218"},{"issue":"2","key":"2488_CR34","doi-asserted-by":"publisher","first-page":"566","DOI":"10.1109\/JSSC.2007.914337","volume":"43","author":"P Lichtsteiner","year":"2008","unstructured":"Lichtsteiner, P., Posch, C., & Delbruck, T. (2008). A $$128 \\times 128$$ 120 db 15$$\\mu $$s latency asynchronous temporal contrast vision sensor. IEEE journal of solid-state circuits (JSSC), 43(2), 566\u2013576.","journal-title":"IEEE journal of solid-state circuits (JSSC)"},{"key":"2488_CR35","doi-asserted-by":"crossref","unstructured":"Liu, H., Qu, S., Lu, F., Bu, Z., Roehrbein, F., Knoll, A., & Chen, G. (2024a) Pcdepth: Pattern-based complementary learning for monocular depth estimation by best of both worlds. arXiv preprint arXiv:2402.18925","DOI":"10.1109\/IROS58592.2024.10802220"},{"key":"2488_CR36","doi-asserted-by":"crossref","unstructured":"Liu, X., Li, J., Zhang, X., Sun, J., Fan, X., & Tian, Y. (2022) Learning visible surface area estimation for irregular objects. In: Proceedings of the 30th ACM International Conference on Multimedia (ACM MM), (pp 2333\u20132343)","DOI":"10.1145\/3503161.3548017"},{"key":"2488_CR37","doi-asserted-by":"crossref","unstructured":"Liu, X., Li, J., Shi, J., Fan, X., Tian, Y., & Zhao, D. (2024b) Event-based monocular depth estimation with recurrent transformers. IEEE Transactions on Circuits and Systems for Video Technology (TCSVT)","DOI":"10.1109\/TCSVT.2024.3378742"},{"key":"2488_CR38","doi-asserted-by":"crossref","unstructured":"Maqueda, A. I., Loquercio, A., Gallego, G., Garc\u00eda, N., & Scaramuzza, D. (2018) Event-based vision meets deep learning on steering prediction for self-driving cars. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 5419\u20135427)","DOI":"10.1109\/CVPR.2018.00568"},{"key":"2488_CR39","doi-asserted-by":"crossref","unstructured":"Mostafavi, M., Yoon, K. J., & Choi, J. (2021) Event-intensity stereo: Estimating depth by the best of both worlds. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), (pp 4258\u20134267)","DOI":"10.1109\/ICCV48922.2021.00422"},{"key":"2488_CR40","doi-asserted-by":"crossref","unstructured":"Pan, T., Cao, Z., & Wang, L. (2023) Srfnet: Monocular depth estimation with fine-grained structure via spatial reliability-oriented fusion of frames and events. arXiv preprint arXiv:2309.12842","DOI":"10.1109\/ICRA57147.2024.10610921"},{"key":"2488_CR41","unstructured":"Paszke, A., Gross, S., Chintala, S., Chanan, G., Yang, E., DeVito, Z., Lin, Z., Desmaison, A., Antiga, L., & Lerer, A. (2017) Automatic differentiation in pytorch. In: Proceedings of the Advances in Neural Information Processing Systems (NeurIPS)"},{"key":"2488_CR42","unstructured":"Prophesee (2021) Prophesee evaluation kit 2 - hd brochure"},{"key":"2488_CR43","doi-asserted-by":"crossref","unstructured":"Ranftl, R., Bochkovskiy, A., & Koltun, V. (2021) Vision transformers for dense prediction. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), (pp 12179\u201312188)","DOI":"10.1109\/ICCV48922.2021.01196"},{"issue":"12","key":"2488_CR44","doi-asserted-by":"publisher","first-page":"1394","DOI":"10.1007\/s11263-017-1050-6","volume":"126","author":"H Rebecq","year":"2018","unstructured":"Rebecq, H., Gallego, G., Mueggler, E., & Scaramuzza, D. (2018). Emvs: Event-based multi-view stereo-3d reconstruction with an event camera in real-time. International Journal of Computer Vision (IJCV), 126(12), 1394\u20131414.","journal-title":"International Journal of Computer Vision (IJCV)"},{"key":"2488_CR45","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., & Brox, T. (2015) U-net: Convolutional networks for biomedical image segmentation. In: Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Springer, (pp 234\u2013241)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"2488_CR46","doi-asserted-by":"crossref","unstructured":"Shang, W., Ren, D., Zou, D., Ren, J. S., Luo, P., & Zuo, W. (2021) Bringing events into video deblurring with non-consecutively blurry frames. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV)","DOI":"10.1109\/ICCV48922.2021.00449"},{"key":"2488_CR47","doi-asserted-by":"crossref","unstructured":"Shi, D., Jing, L., Li, R., Liu, Z., Wang, L., Xu, H., & Zhang, Y. (2023a) Improved event-based dense depth estimation via optical flow compensation. In: Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), IEEE, (pp 4902\u20134908)","DOI":"10.1109\/ICRA48891.2023.10160605"},{"key":"2488_CR48","doi-asserted-by":"crossref","unstructured":"Shi, P., Peng, J., Qiu, J., Ju, X., Lo, F. P. W., & Lo, B. (2023b) Even: An event-based framework for monocular depth estimation at adverse night conditions. In: Proceedings of the IEEE International Conference on Robotics and Biomimetics (ROBIO), IEEE, (pp 1\u20137)","DOI":"10.1109\/ROBIO58561.2023.10354658"},{"key":"2488_CR49","unstructured":"Shi, X., Chen, Z., Wang, H., Yeung, D. Y., Wong, W. K., & Woo, W. c. (2015) Convolutional lstm network: A machine learning approach for precipitation nowcasting. Proceedings of the Advances in Neural Information Processing Systems (NeurIPS) 28"},{"key":"2488_CR50","first-page":"369","volume":"11006","author":"LN Smith","year":"2019","unstructured":"Smith, L. N., & Topin, N. (2019). Super-convergence: Very fast training of neural networks using large learning rates. Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications, SPIE, 11006, 369\u2013386.","journal-title":"Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications, SPIE"},{"key":"2488_CR51","doi-asserted-by":"crossref","unstructured":"Su, Q., Chou, Y., Hu, Y., Li, J., Mei, S., Zhang, Z., & Li, G. (2023) Deep directly-trained spiking neural networks for object detection. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), (pp 6555\u20136565)","DOI":"10.1109\/ICCV51070.2023.00603"},{"key":"2488_CR52","doi-asserted-by":"crossref","unstructured":"Tulyakov, S., Gehrig, D., Georgoulis, S., Erbach, J., Gehrig, M., Li ,Y., S & caramuzza, D. (2021) Time lens: Event-based video frame interpolation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 16155\u201316164)","DOI":"10.1109\/CVPR46437.2021.01589"},{"issue":"6","key":"2488_CR53","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3272127.3275041","volume":"37","author":"J Valentin","year":"2018","unstructured":"Valentin, J., Kowdle, A., Barron, J. T., Wadhwa, N., Dzitsiuk, M., Schoenberg, M., Verma, V., Csaszar, A., Turner, E., Dryanovski, I., et al. (2018). Depth from motion for smartphone ar. ACM Transactions on Graphics (ToG), 37(6), 1\u201319.","journal-title":"ACM Transactions on Graphics (ToG)"},{"issue":"2","key":"2488_CR54","doi-asserted-by":"publisher","first-page":"994","DOI":"10.1109\/LRA.2018.2793357","volume":"3","author":"AR Vidal","year":"2018","unstructured":"Vidal, A. R., Rebecq, H., Horstschaefer, T., & Scaramuzza, D. (2018). Ultimate slam? combining events, images, and imu for robust visual slam in hdr and high-speed scenarios. IEEE Robotics and Automation Letters (RAL), 3(2), 994\u20131001.","journal-title":"IEEE Robotics and Automation Letters (RAL)"},{"key":"2488_CR55","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J. Y., & Kweon, I. S. (2018) Cbam: Convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), (pp 3\u201319)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"2488_CR56","doi-asserted-by":"publisher","first-page":"331","DOI":"10.3389\/fnins.2018.00331","volume":"12","author":"Y Wu","year":"2018","unstructured":"Wu, Y., Deng, L., Li, G., Zhu, J., & Shi, L. (2018). Spatio-temporal backpropagation for training high-performance spiking neural networks. Frontiers in Neuroscience (FN), 12, 331.","journal-title":"Frontiers in Neuroscience (FN)"},{"issue":"7","key":"2488_CR57","doi-asserted-by":"publisher","first-page":"2401","DOI":"10.1007\/s11263-023-01979-4","volume":"132","author":"K Xian","year":"2024","unstructured":"Xian, K., Cao, Z., Shen, C., & Lin, G. (2024). Towards robust monocular depth estimation: A new baseline and benchmark. International Journal of Computer Vision (IJCV), 132(7), 2401\u20132419.","journal-title":"International Journal of Computer Vision (IJCV)"},{"issue":"4","key":"2488_CR58","doi-asserted-by":"publisher","first-page":"1012","DOI":"10.1007\/s11263-023-01915-6","volume":"132","author":"M Xiang","year":"2024","unstructured":"Xiang, M., Dai, Y., Zhang, F., Shi, J., Tian, X., & Zhang, Z. (2024). Towards a unified network for robust monocular depth estimation: Network architecture, training strategy and dataset. International Journal of Computer Vision (IJCV), 132(4), 1012\u20131028.","journal-title":"International Journal of Computer Vision (IJCV)"},{"key":"2488_CR59","doi-asserted-by":"crossref","unstructured":"Yang, Y., Han, J., Liang, J., Sato, I., & Shi, B. (2023) Learning event guided high dynamic range video reconstruction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 13924\u201313934)","DOI":"10.1109\/CVPR52729.2023.01338"},{"key":"2488_CR60","doi-asserted-by":"crossref","unstructured":"Yuan, W., Gu, X., Dai, Z., Zhu, S., & Tan, P. (2022) Neural window fully-connected crfs for monocular depth estimation.In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 3916\u20133925)","DOI":"10.1109\/CVPR52688.2022.00389"},{"key":"2488_CR61","doi-asserted-by":"crossref","unstructured":"Yucel, M. K., Dimaridou, V., Drosou, A., & Saa-Garriga, A. (2021) Real-time monocular depth estimation with sparse supervision on mobile. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), (pp 2428\u20132437)","DOI":"10.1109\/CVPRW53098.2021.00275"},{"key":"2488_CR62","doi-asserted-by":"crossref","unstructured":"Zhang, J., Yang, X., Fu, Y., Wei, X., Yin, B., & Dong, B. (2021) Object tracking by jointly exploiting frame and event domain. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV)","DOI":"10.1109\/ICCV48922.2021.01280"},{"key":"2488_CR63","doi-asserted-by":"crossref","unstructured":"Zhang, J., Wang, Y., Liu, W., Li, M., Bai, J., Yin, B., & Yang, X. (2023a) Frame-event alignment and fusion network for high frame rate tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 9781\u20139790)","DOI":"10.1109\/CVPR52729.2023.00943"},{"key":"2488_CR64","doi-asserted-by":"crossref","unstructured":"Zhang, J., Dong, B., Fu, Y., Wang, Y., Wei, X., Yin, B., & Yang, X. (2024). A universal event-based plug-in module for visual object tracking in degraded conditions. International Journal of Computer Vision (IJCV), 132(5), 1857\u20131879.","DOI":"10.1007\/s11263-023-01959-8"},{"key":"2488_CR65","doi-asserted-by":"crossref","unstructured":"Zhang, N., Nex, F., Vosselman, G., & Kerle, N. (2023b) Lite-mono: A lightweight cnn and transformer architecture for self-supervised monocular depth estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 18537\u201318546)","DOI":"10.1109\/CVPR52729.2023.01778"},{"issue":"3","key":"2488_CR66","doi-asserted-by":"publisher","first-page":"2032","DOI":"10.1109\/LRA.2018.2800793","volume":"3","author":"AZ Zhu","year":"2018","unstructured":"Zhu, A. Z., Thakur, D., \u00d6zaslan, T., Pfrommer, B., Kumar, V., & Daniilidis, K. (2018). The multivehicle stereo event camera dataset: An event camera dataset for 3d perception. IEEE Robotics and Automation Letters (RAL), 3(3), 2032\u20132039.","journal-title":"IEEE Robotics and Automation Letters (RAL)"},{"key":"2488_CR67","doi-asserted-by":"crossref","unstructured":"Zhu, A. Z., Yuan, L., Chaney, K., & Daniilidis, K. (2019) Unsupervised event-based learning of optical flow, depth, and egomotion. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (pp 989\u2013997)","DOI":"10.1109\/CVPR.2019.00108"},{"key":"2488_CR68","doi-asserted-by":"crossref","unstructured":"Zhu, J., Liu, L., Jiang, B., Wen, F., Zhang, H., Li, W., & Liu, Y. (2023) Self-supervised event-based monocular depth estimation using cross-modal consistency. In: Proceedings of the IEEE International Conference on Intelligent Robots and Systems (IROS), (pp 7704\u20137710)","DOI":"10.1109\/IROS55552.2023.10342434"},{"key":"2488_CR69","doi-asserted-by":"crossref","unstructured":"Zhu, L., Li, J., Wang, X., Huang, T., & Tian, Y. (2021) Neuspike-net: High speed video reconstruction via bio-inspired neuromorphic cameras. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV)","DOI":"10.1109\/ICCV48922.2021.00240"},{"key":"2488_CR70","doi-asserted-by":"crossref","unstructured":"Zuo, Y. F., Cui, L., Peng, X., Xu, Y., Gao, S., Wang, X., & Kneip, L. (2021) Accurate depth estimation from a hybrid event-rgb stereo setup. In: Proceedings of the IEEE International Conference on Intelligent Robots and Systems (IROS), (pp 6833\u20136840)","DOI":"10.1109\/IROS51168.2021.9635834"},{"key":"2488_CR71","doi-asserted-by":"crossref","unstructured":"Zuo, Y. F., Yang, J., Chen, J., Wang, X., Wang, Y., & Kneip, L. (2022) Devo: Depth-event camera visual odometry in challenging conditions. In: Proceedings of the IEEE International Conference on Robotics and Automation (ICRA) (pp 2179\u20132185)","DOI":"10.1109\/ICRA46639.2022.9811805"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02488-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-025-02488-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02488-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T08:54:26Z","timestamp":1760086466000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-025-02488-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,22]]},"references-count":71,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["2488"],"URL":"https:\/\/doi.org\/10.1007\/s11263-025-02488-2","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,7,22]]},"assertion":[{"value":"17 July 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 May 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 July 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no conflict of interest to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This article does not contain any studies with human participants performed by any of the authors.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics Approval"}},{"value":"Informed consent was obtained from all individual participants included in the study.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to Participate"}},{"value":"Consent for publication was obtained from the participants.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for Publication"}}]}}