{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T17:45:31Z","timestamp":1772905531182,"version":"3.50.1"},"publisher-location":"Cham","reference-count":37,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031734632","type":"print"},{"value":"9783031734649","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,12,4]],"date-time":"2024-12-04T00:00:00Z","timestamp":1733270400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,4]],"date-time":"2024-12-04T00:00:00Z","timestamp":1733270400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73464-9_12","type":"book-chapter","created":{"date-parts":[[2024,12,3]],"date-time":"2024-12-03T10:02:57Z","timestamp":1733220177000},"page":"190-205","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Edge-Guided Fusion and\u00a0Motion Augmentation for\u00a0Event-Image Stereo"],"prefix":"10.1007","author":[{"given":"Fengan","family":"Zhao","sequence":"first","affiliation":[]},{"given":"Qianang","family":"Zhou","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0128-4960","authenticated-orcid":false,"given":"Junlin","family":"Xiong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,4]]},"reference":[{"key":"12_CR1","doi-asserted-by":"crossref","unstructured":"Ahmed, S.H., Jang, H.W., Uddin, S.N., Jung, Y.J.: Deep event stereo leveraged by event-to-image translation. In: AAAI, pp. 882\u2013890 (2021)","DOI":"10.1609\/aaai.v35i2.16171"},{"key":"12_CR2","unstructured":"Ba, J.L., Kiros, J.R., Hinton, G.E.: Layer normalization. arXiv preprint arXiv:1607.06450 (2016)"},{"issue":"10","key":"12_CR3","doi-asserted-by":"publisher","first-page":"2333","DOI":"10.1109\/JSSC.2014.2342715","volume":"49","author":"C Brandli","year":"2014","unstructured":"Brandli, C., Berner, R., Yang, M., Liu, S.C., Delbruck, T.: A 240 $$\\times $$ 180 130 dB 3 $$\\mu $$s latency global shutter spatiotemporal vision sensor. IEEE J. Solid-State Circuits 49(10), 2333\u20132341 (2014)","journal-title":"IEEE J. Solid-State Circuits"},{"key":"12_CR4","doi-asserted-by":"crossref","unstructured":"Chang, J.R., Chen, Y.S.: Pyramid stereo matching network. In: CVPR, pp. 5410\u20135418 (2018)","DOI":"10.1109\/CVPR.2018.00567"},{"key":"12_CR5","first-page":"22158","volume":"33","author":"X Cheng","year":"2020","unstructured":"Cheng, X., et al.: Hierarchical neural architecture search for deep stereo matching. Adv. Neural. Inf. Process. Syst. 33, 22158\u201322169 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"12_CR6","doi-asserted-by":"crossref","unstructured":"Cho, H., Yoon, K.J.: Event-image fusion stereo using cross-modality feature propagation. In: AAAI, pp. 454\u2013462 (2022)","DOI":"10.1609\/aaai.v36i1.19923"},{"key":"12_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"470","DOI":"10.1007\/978-3-031-19824-3_28","volume-title":"Computer Vision \u2013 ECCV 2022","author":"H Cho","year":"2022","unstructured":"Cho, H., Yoon, K.J.: Selection and cross similarity for event-image deep stereo. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 470\u2013486. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19824-3_28"},{"issue":"1","key":"12_CR8","doi-asserted-by":"publisher","first-page":"154","DOI":"10.1109\/TPAMI.2020.3008413","volume":"44","author":"G Gallego","year":"2020","unstructured":"Gallego, G., et al.: Event-based vision: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 154\u2013180 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"12_CR9","doi-asserted-by":"publisher","first-page":"122743","DOI":"10.1016\/j.eswa.2023.122743","volume":"241","author":"DK Ghosh","year":"2024","unstructured":"Ghosh, D.K., Jung, Y.J.: Two-stage cross-fusion network for stereo event-based depth estimation. Expert Syst. Appl. 241, 122743 (2024). https:\/\/doi.org\/10.1016\/j.eswa.2023.122743","journal-title":"Expert Syst. Appl."},{"key":"12_CR10","doi-asserted-by":"crossref","unstructured":"Guo, X., Yang, K., Yang, W., Wang, X., Li, H.: Group-wise correlation stereo network. In: CVPR, pp. 3273\u20133282 (2019)","DOI":"10.1109\/CVPR.2019.00339"},{"key":"12_CR11","unstructured":"Huang, K., Zhang, S., Zhang, J., Tao, D.: Event-based simultaneous localization and mapping: a comprehensive survey. arXiv preprint arXiv:2304.09793 (2023)"},{"key":"12_CR12","unstructured":"Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International Conference on Machine Learning, pp. 448\u2013456 (2015)"},{"key":"12_CR13","doi-asserted-by":"crossref","unstructured":"Jiang, S., Campbell, D., Lu, Y., Li, H., Hartley, R.: Learning to estimate hidden motions with global motion aggregation. In: ICCV, pp. 9772\u20139781 (2021)","DOI":"10.1109\/ICCV48922.2021.00963"},{"key":"12_CR14","doi-asserted-by":"crossref","unstructured":"Kendall, A., et al.: End-to-end learning of geometry and context for deep stereo regression. In: ICCV, pp. 66\u201375 (2017)","DOI":"10.1109\/ICCV.2017.17"},{"key":"12_CR15","doi-asserted-by":"crossref","unstructured":"Li, J., et al.: Practical stereo matching via cascaded recurrent network with adaptive correlation. In: CVPR, pp. 16263\u201316272 (2022)","DOI":"10.1109\/CVPR52688.2022.01578"},{"key":"12_CR16","doi-asserted-by":"crossref","unstructured":"Li, W., Zhu, X., Gong, S.: Harmonious attention network for person re-identification. In: CVPR, pp. 2285\u20132294 (2018)","DOI":"10.1109\/CVPR.2018.00243"},{"issue":"2","key":"12_CR17","doi-asserted-by":"publisher","first-page":"566","DOI":"10.1109\/JSSC.2007.914337","volume":"43","author":"P Lichtsteiner","year":"2008","unstructured":"Lichtsteiner, P., Posch, C., Delbruck, T.: A 128 $$\\times $$ 128 120 dB 15 $$\\mu $$s latency asynchronous temporal contrast vision sensor. IEEE J. Solid-State Circuits 43(2), 566\u2013576 (2008)","journal-title":"IEEE J. Solid-State Circuits"},{"key":"12_CR18","doi-asserted-by":"crossref","unstructured":"Lipson, L., Teed, Z., Deng, J.: Raft-stereo: multilevel recurrent field transforms for stereo matching. In: 2021 International Conference on 3D Vision (3DV), pp. 218\u2013227 (2021)","DOI":"10.1109\/3DV53792.2021.00032"},{"key":"12_CR19","doi-asserted-by":"crossref","unstructured":"Liu, H., et al.: TMA: temporal motion aggregation for event-based optical flow. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00888"},{"key":"12_CR20","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"issue":"4","key":"12_CR21","doi-asserted-by":"publisher","first-page":"900","DOI":"10.1007\/s11263-020-01410-2","volume":"129","author":"M Mostafavi","year":"2021","unstructured":"Mostafavi, M., Wang, L., Yoon, K.J.: Learning to reconstruct HDR images from events, with applications to depth and flow prediction. Int. J. Comput. Vision 129(4), 900\u2013920 (2021)","journal-title":"Int. J. Comput. Vision"},{"key":"12_CR22","doi-asserted-by":"publisher","unstructured":"Mostafavi\u00a0I, S.M., Yoon, K.J., Choi, J.: Event-intensity stereo: estimating depth by the best of both worlds. In: ICCV, pp. 4238\u20134247 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.00422","DOI":"10.1109\/ICCV48922.2021.00422"},{"key":"12_CR23","doi-asserted-by":"crossref","unstructured":"Park, T., Liu, M.Y., Wang, T.C., Zhu, J.Y.: Semantic image synthesis with spatially-adaptive normalization. In: CVPR, pp. 2332\u20132341 (2019)","DOI":"10.1109\/CVPR.2019.00244"},{"key":"12_CR24","unstructured":"Paszke, A., et\u00a0al.: PyTorch: an imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 32 (2019)"},{"key":"12_CR25","doi-asserted-by":"crossref","unstructured":"Smith, L.N., Topin, N.: Super-convergence: very fast training of neural networks using large learning rates. In: Artificial Intelligence and Machine Learning for Multi-domain Operations Applications, vol. 11006, pp. 369\u2013386 (2019)","DOI":"10.1117\/12.2520589"},{"key":"12_CR26","doi-asserted-by":"crossref","unstructured":"Soria, X., Li, Y., Rouhani, M., Sappa, A.D.: Tiny and efficient model for the edge detection generalization. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) Workshops, pp. 1364\u20131373 (2023)","DOI":"10.1109\/ICCVW60793.2023.00147"},{"key":"12_CR27","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"412","DOI":"10.1007\/978-3-031-19797-0_24","volume-title":"Computer Vision \u2013 ECCV 2022","author":"L Sun","year":"2022","unstructured":"Sun, L., et al.: Event-based fusion for motion deblurring with cross-modal attention. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13678, pp. 412\u2013428. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19797-0_24"},{"key":"12_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"402","DOI":"10.1007\/978-3-030-58536-5_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Z Teed","year":"2020","unstructured":"Teed, Z., Deng, J.: RAFT: recurrent all-pairs field transforms for optical flow. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12347, pp. 402\u2013419. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_24"},{"key":"12_CR29","doi-asserted-by":"crossref","unstructured":"Tulyakov, S., Fleuret, F., Kiefel, M., Gehler, P., Hirsch, M.: Learning an event sequence embedding for dense event-based deep stereo. In: ICCV, pp. 1527\u20131537 (2019)","DOI":"10.1109\/ICCV.2019.00161"},{"key":"12_CR30","doi-asserted-by":"crossref","unstructured":"Wang, F., Galliani, S., Vogel, C., Pollefeys, M.: IterMVS: iterative probability estimation for efficient multi-view stereo. In: CVPR, pp. 8606\u20138615 (2022)","DOI":"10.1109\/CVPR52688.2022.00841"},{"key":"12_CR31","doi-asserted-by":"crossref","unstructured":"Xu, G., Wang, X., Ding, X., Yang, X.: Iterative geometry encoding volume for stereo matching. In: CVPR, pp. 21919\u201321928 (2023)","DOI":"10.1109\/CVPR52729.2023.02099"},{"issue":"11","key":"12_CR32","doi-asserted-by":"publisher","first-page":"13941","DOI":"10.1109\/TPAMI.2023.3298645","volume":"45","author":"H Xu","year":"2023","unstructured":"Xu, H., et al.: Unifying flow, stereo and depth estimation. IEEE Trans. Pattern Anal. Mach. Intell. 45(11), 13941\u201313958 (2023). https:\/\/doi.org\/10.1109\/TPAMI.2023.3298645","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"12_CR33","doi-asserted-by":"crossref","unstructured":"Xu, H., Zhang, J.: AANet: adaptive aggregation network for efficient stereo matching. In: CVPR, pp. 1959\u20131968 (2020)","DOI":"10.1109\/CVPR42600.2020.00203"},{"key":"12_CR34","doi-asserted-by":"crossref","unstructured":"Zhang, F., Prisacariu, V., Yang, R., Torr, P.H.: GA-net: guided aggregation net for end-to-end stereo matching. In: CVPR, pp. 185\u2013194 (2019)","DOI":"10.1109\/CVPR.2019.00027"},{"key":"12_CR35","doi-asserted-by":"publisher","unstructured":"Zhang, K., et al.: Discrete time convolution for fast event-based stereo. In: CVPR, pp. 8666\u20138676 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.00848","DOI":"10.1109\/CVPR52688.2022.00848"},{"issue":"3","key":"12_CR36","doi-asserted-by":"publisher","first-page":"2032","DOI":"10.1109\/LRA.2018.2800793","volume":"3","author":"AZ Zhu","year":"2018","unstructured":"Zhu, A.Z., Thakur, D., \u00d6zaslan, T., Pfrommer, B., Kumar, V., Daniilidis, K.: The multivehicle stereo event camera dataset: an event camera dataset for 3D perception. IEEE Robot. Autom. Lett. 3(3), 2032\u20132039 (2018). https:\/\/doi.org\/10.1109\/LRA.2018.2800793","journal-title":"IEEE Robot. Autom. Lett."},{"key":"12_CR37","doi-asserted-by":"crossref","unstructured":"Zhu, A.Z., Yuan, L., Chaney, K., Daniilidis, K.: Unsupervised event-based learning of optical flow, depth, and egomotion. In: CVPR, pp. 989\u2013997 (2019)","DOI":"10.1109\/CVPR.2019.00108"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73464-9_12","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,3]],"date-time":"2024-12-03T10:06:44Z","timestamp":1733220404000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73464-9_12"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,4]]},"ISBN":["9783031734632","9783031734649"],"references-count":37,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73464-9_12","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,4]]},"assertion":[{"value":"4 December 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}