{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T19:43:56Z","timestamp":1768074236402,"version":"3.49.0"},"publisher-location":"Cham","reference-count":47,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031783531","type":"print"},{"value":"9783031783548","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,12,4]],"date-time":"2024-12-04T00:00:00Z","timestamp":1733270400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,4]],"date-time":"2024-12-04T00:00:00Z","timestamp":1733270400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-78354-8_30","type":"book-chapter","created":{"date-parts":[[2024,12,3]],"date-time":"2024-12-03T10:27:42Z","timestamp":1733221662000},"page":"475-491","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["SDformerFlow: Spiking Neural Network Transformer for Event-based Optical Flow"],"prefix":"10.1007","author":[{"given":"Yi","family":"Tian","sequence":"first","affiliation":[]},{"given":"Juan","family":"Andrade-Cetto","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,4]]},"reference":[{"key":"30_CR1","doi-asserted-by":"crossref","unstructured":"de\u00a0Blegiers, T., Dave, I.R., et\u00a0al.: Eventtransact: A video transformer-based framework for event-camera based action recognition. In: IEEE\/RSJ Int. Conf. Intell. Robots Syst. pp. 1261\u20131267 (2023)","DOI":"10.1109\/IROS55552.2023.10341740"},{"key":"30_CR2","doi-asserted-by":"publisher","first-page":"1160034","DOI":"10.3389\/fnins.2023.1160034","volume":"17","author":"J Cuadrado","year":"2023","unstructured":"Cuadrado, J., Ran\u00e7on, U., et al.: Optical flow estimation from event-based cameras and spiking neural networks. Front. Neurosci. 17, 1160034 (2023)","journal-title":"Front. Neurosci."},{"key":"30_CR3","doi-asserted-by":"crossref","unstructured":"Ding, Z., Zhao, R., et\u00a0al.: Spatio-temporal recurrent networks for event-based optical flow estimation. In: AAAI Conf. Artif. Intell. vol.\u00a036, pp. 525\u2013533 (2021)","DOI":"10.1609\/aaai.v36i1.19931"},{"key":"30_CR4","unstructured":"Dosovitskiy, A., Beyer, L., et\u00a0al.: An image is worth 16x16 words: Transformers for image recognition at scale. In: Int. Conf. Learn. Represent. (2020)"},{"key":"30_CR5","unstructured":"Fan, Q., You, Q., et\u00a0al.: Vitar: Vision transformer with any resolution. arXiv preprint arXiv:2403.18361 (2024)"},{"key":"30_CR6","doi-asserted-by":"crossref","unstructured":"Fang, W., Chen, Y., et\u00a0al.: Spikingjelly: An open-source machine learning infrastructure platform for spike-based intelligence. Sci. Adv. 9(40), eadi1480 (2023)","DOI":"10.1126\/sciadv.adi1480"},{"key":"30_CR7","unstructured":"Fang, W., Yu, Z., et\u00a0al.: Deep residual learning in spiking neural networks. In: Conf. Neural Inf. Process. Syst. vol.\u00a034, pp. 21056\u201321069 (2021)"},{"issue":"12","key":"30_CR8","doi-asserted-by":"publisher","first-page":"14081","DOI":"10.1109\/TPAMI.2023.3300741","volume":"45","author":"Y Gao","year":"2023","unstructured":"Gao, Y., Lu, J., et al.: Action recognition and benchmark using event cameras. IEEE Trans. Pattern Anal. Mach. Intell. 45(12), 14081\u201314097 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"30_CR9","doi-asserted-by":"publisher","first-page":"4947","DOI":"10.1109\/LRA.2021.3068942","volume":"6","author":"M Gehrig","year":"2021","unstructured":"Gehrig, M., Aarents, W., et al.: DSEC: A stereo event camera dataset for driving scenarios. IEEE Robotics Autom. Lett. 6(3), 4947\u20134954 (2021)","journal-title":"IEEE Robotics Autom. Lett."},{"key":"30_CR10","doi-asserted-by":"crossref","unstructured":"Gehrig, M., Millhausler, M., et\u00a0al.: E-RAFT: Dense optical flow from event cameras. In: Int. Conf. 3D Vis. pp. 197\u2013206 (2021)","DOI":"10.1109\/3DV53792.2021.00030"},{"key":"30_CR11","doi-asserted-by":"crossref","unstructured":"Guizilini, V., Ambrus, R., et\u00a0al.: Multi-frame self-supervised depth with transformers. In: IEEE\/CVF Conf. Comput. Vis. Pattern Recognit. pp. 160\u2013170 (2022)","DOI":"10.1109\/CVPR52688.2022.00026"},{"key":"30_CR12","unstructured":"Hagenaars, J., Paredes-Vall\u00e9s, F., de\u00a0Croon, G.: Self-supervised learning of event-based optical flow with spiking neural networks. In: Conf. Neural Inf. Process. Syst. vol.\u00a034 (2021)"},{"key":"30_CR13","doi-asserted-by":"crossref","unstructured":"Horowitz, M.: 1.1 computing\u2019s energy problem (and what we can do about it). In: IEEE Int. Solid-State Circuits Conf. pp. 10\u201314 (2014)","DOI":"10.1109\/ISSCC.2014.6757323"},{"key":"30_CR14","doi-asserted-by":"crossref","unstructured":"Hu, Y., Deng, L., et\u00a0al.: Advancing spiking neural networks toward deep residual learning. IEEE Trans. Neural Networks Learn. Syst. (2024), early access","DOI":"10.1109\/TNNLS.2024.3355393"},{"key":"30_CR15","doi-asserted-by":"crossref","unstructured":"Huang, Z., Shi, X., et\u00a0al.: Flowformer: A transformer architecture for optical flow. In: Eur. Conf. Comput. Vis. pp. 668\u2013685 (2022)","DOI":"10.1007\/978-3-031-19790-1_40"},{"key":"30_CR16","doi-asserted-by":"crossref","unstructured":"Kosta, A.K., Roy, K.: Adaptive-spikenet: Event-based optical flow estimation using spiking neural networks with learnable neuronal dynamics. In: IEEE Int. Conf. Robotics Autom. pp. 6021\u20136027 (2023)","DOI":"10.1109\/ICRA48891.2023.10160551"},{"key":"30_CR17","doi-asserted-by":"crossref","unstructured":"Lee, C., Kosta, A., et\u00a0al.: Spike-flownet: Event-based optical flow estimation with energy-efficient hybrid neural networks. In: Eur. Conf. Comput. Vis. pp. 366\u2013382 (2020)","DOI":"10.1007\/978-3-030-58526-6_22"},{"key":"30_CR18","doi-asserted-by":"crossref","unstructured":"Li, Y., Huang, Z., et\u00a0al.: Blinkflow: A dataset to push the limits of event-based optical flow estimation. In: IEEE\/RSJ Int. Conf. Intell. Robots Syst. pp. 3881\u20133888 (2023)","DOI":"10.1109\/IROS55552.2023.10341802"},{"key":"30_CR19","doi-asserted-by":"crossref","unstructured":"Liu, H., Chen, G., et\u00a0al.: TMA: Temporal motion aggregation for event-based optical flow. In: IEEE Int. Conf. Comput. Vis. pp. 9651\u20139660 (2023)","DOI":"10.1109\/ICCV51070.2023.00888"},{"key":"30_CR20","doi-asserted-by":"crossref","unstructured":"Liu, Z., Hu, H., et\u00a0al.: Swin transformer v2: Scaling up capacity and resolution. In: IEEE\/CVF Conf. Comput. Vis. Pattern Recognit. pp. 12009\u201312019 (2022)","DOI":"10.1109\/CVPR52688.2022.01170"},{"key":"30_CR21","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., et\u00a0al.: Swin transformer: Hierarchical vision transformer using shifted windows. In: IEEE Int. Conf. Comput. Vis. pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"30_CR22","doi-asserted-by":"crossref","unstructured":"Liu, Z., Ning, J., et\u00a0al.: Video swin transformer. In: IEEE\/CVF Conf. Comput. Vis. Pattern Recognit. pp. 3202\u20133211 (2022)","DOI":"10.1109\/CVPR52688.2022.00320"},{"key":"30_CR23","doi-asserted-by":"crossref","unstructured":"Lu, Y., Wang, Q., et\u00a0al.: Transflow: Transformer as flow learner. In: IEEE\/CVF Conf. Comput. Vis. Pattern Recognit. pp. 18063\u201318073 (2023)","DOI":"10.1109\/CVPR52729.2023.01732"},{"key":"30_CR24","doi-asserted-by":"crossref","unstructured":"Luo, X., Luo, K., et\u00a0al.: Learning optical flow from event camera with rendered dataset. arXiv preprint arXiv:2303.11011 (2023)","DOI":"10.1109\/ICCV51070.2023.00903"},{"issue":"6","key":"30_CR25","doi-asserted-by":"publisher","first-page":"51","DOI":"10.1109\/MSP.2019.2931595","volume":"36","author":"EO Neftci","year":"2019","unstructured":"Neftci, E.O., Mostafa, H., Zenke, F.: Surrogate gradient learning in spiking neural networks: Bringing the power of gradient-based optimization to spiking neural networks. IEEE Signal Process. Mag. 36(6), 51\u201363 (2019)","journal-title":"IEEE Signal Process. Mag."},{"key":"30_CR26","doi-asserted-by":"crossref","unstructured":"Paredes-Vall\u00e9s, F., Scheper, K.Y.W., et\u00a0al.: Taming contrast maximization for learning sequential, low-latency, event-based optical flow. In: IEEE Int. Conf. Comput. Vis. pp. 9695\u20139705 (2023)","DOI":"10.1109\/ICCV51070.2023.00889"},{"key":"30_CR27","doi-asserted-by":"crossref","unstructured":"Ponghiran, W., Liyanagedera, C.M., Roy, K.: Event-based temporally dense optical flow estimation with sequential learning. In: IEEE Int. Conf. Comput. Vis. pp. 9827\u20139836 (2023)","DOI":"10.1109\/ICCV51070.2023.00901"},{"key":"30_CR28","doi-asserted-by":"crossref","unstructured":"Shiba, S., Aoki, Y., Gallego, G.: Secrets of event-based optical flow. In: Eur. Conf. Comput. Vis. pp. 628\u2013645 (2022)","DOI":"10.1007\/978-3-031-19797-0_36"},{"key":"30_CR29","doi-asserted-by":"crossref","unstructured":"Sui, X., Li, S., et\u00a0al.: CRAFT: Cross-attentional flow transformer for robust optical flow. In: IEEE\/CVF Conf. Comput. Vis. Pattern Recognit. pp. 17581\u20131790 (2022)","DOI":"10.1109\/CVPR52688.2022.01708"},{"key":"30_CR30","doi-asserted-by":"crossref","unstructured":"Teed, Z., Deng, J.: RAFT: Recurrent all-pairs field transforms for optical flow. In: Eur. Conf. Comput. Vis. pp. 402\u2013419 (2020)","DOI":"10.1007\/978-3-030-58536-5_24"},{"key":"30_CR31","doi-asserted-by":"crossref","unstructured":"Tian, R., Wu, Z., et\u00a0al.: Resformer: Scaling vits with multi-resolution training. In: IEEE\/CVF Conf. Comput. Vis. Pattern Recognit. pp. 22721\u201322731 (2023)","DOI":"10.1109\/CVPR52729.2023.02176"},{"key":"30_CR32","unstructured":"Tian, Y., Andrade-Cetto, J.: Event transformer flownet for optical flow estimation. In: British Mach. Vis. Conf. (2022)"},{"key":"30_CR33","doi-asserted-by":"crossref","unstructured":"Wang, Y., Shi, K., et\u00a0al.: Spatial-temporal self-attention for asynchronous spiking neural networks. In: Int. Joint Conf. Artif. Intell. pp. 3085\u20133093 (2023)","DOI":"10.24963\/ijcai.2023\/344"},{"key":"30_CR34","doi-asserted-by":"crossref","unstructured":"Wang, Z., Fang, Y., et\u00a0al.: Masked spiking transformer. In: IEEE Int. Conf. Comput. Vis. pp. 1761\u20131771 (2023)","DOI":"10.1109\/ICCV51070.2023.00169"},{"key":"30_CR35","doi-asserted-by":"crossref","unstructured":"Wu, Y., Paredes-Vall\u00e9s, F., de\u00a0Croon, G.C.H.E.: Rethinking event-based optical flow: Iterative deblurring as an alternative to correlation volumes. arXiv preprint arXiv:2211.13726 (2023)","DOI":"10.1109\/ICRA57147.2024.10610353"},{"key":"30_CR36","doi-asserted-by":"crossref","unstructured":"Xu, H., Zhang, J., et\u00a0al.: Gmflow: Learning optical flow via global matching. In: IEEE\/CVF Conf. Comput. Vis. Pattern Recognit. pp. 8121\u20138130 (2022)","DOI":"10.1109\/CVPR52688.2022.00795"},{"key":"30_CR37","doi-asserted-by":"crossref","unstructured":"Yang, Y., Pan, L., Liu, L.: Event camera data pre-training. arXiv preprint arXiv:2301.01928 (2023)","DOI":"10.1109\/ICCV51070.2023.00982"},{"key":"30_CR38","unstructured":"Yao, M., Hu, J., et\u00a0al.: Spike-driven transformer v2: Meta spiking neural network architecture inspiring the design of next-generation neuromorphic chips. In: Int. Conf. Learn. Represent. (2024)"},{"key":"30_CR39","unstructured":"Yao, M., Hu, J., et\u00a0al.: Spike-driven transformer. In: Conf. Neural Inf. Process. Syst. (2023)"},{"issue":"8","key":"30_CR40","doi-asserted-by":"publisher","first-page":"9393","DOI":"10.1109\/TPAMI.2023.3241201","volume":"45","author":"M Yao","year":"2023","unstructured":"Yao, M., Zhao, G., et al.: Attention spiking neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 45(8), 9393\u20139410 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"2","key":"30_CR41","doi-asserted-by":"publisher","first-page":"896","DOI":"10.1109\/TPAMI.2023.3329173","volume":"46","author":"W Yu","year":"2024","unstructured":"Yu, W., Si, C., et al.: Metaformer baselines for vision. IEEE Trans. Pattern Anal. Mach. Intell. 46(2), 896\u2013912 (2024)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"30_CR42","doi-asserted-by":"crossref","unstructured":"Zhang, J., Tang, L., et\u00a0al.: Spike transformer: Monocular depth estimation for spiking camera. In: Eur. Conf. Comput. Vis. pp. 34\u201352 (2022)","DOI":"10.1007\/978-3-031-20071-7_3"},{"key":"30_CR43","unstructured":"Zhou, Z., Zhu, Y., et\u00a0al.: Spikformer: When spiking neural network meets transformer. In: Int. Conf. Learn. Represent. (2023)"},{"key":"30_CR44","doi-asserted-by":"crossref","unstructured":"Zhu, A., Yuan, L., et\u00a0al.: EV-FlowNet: Self-supervised optical flow estimation for event-based cameras. In: Robotics Sci. Syst. Conf. (2018)","DOI":"10.15607\/RSS.2018.XIV.062"},{"issue":"3","key":"30_CR45","doi-asserted-by":"publisher","first-page":"2032","DOI":"10.1109\/LRA.2018.2800793","volume":"3","author":"AZ Zhu","year":"2018","unstructured":"Zhu, A.Z., Thakur, D., et al.: The multivehicle stereo event camera dataset: An event camera dataset for 3D perception. IEEE Robotics and Automation Letters 3(3), 2032\u20132039 (2018)","journal-title":"IEEE Robotics and Automation Letters"},{"key":"30_CR46","doi-asserted-by":"crossref","unstructured":"Zhu, A.Z., Yuan, L., et\u00a0al.: Unsupervised event-based learning of optical flow, depth, and egomotion. In: IEEE\/CVF Conf. Comput. Vis. Pattern Recognit. pp. 989\u2013997 (2019)","DOI":"10.1109\/CVPR.2019.00108"},{"key":"30_CR47","unstructured":"Zou, S., Mu, Y., et\u00a0al.: Event-based human pose tracking by spiking spatiotemporal transformer. arXiv preprint arXiv:2303.09681 (2023)"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-78354-8_30","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,3]],"date-time":"2024-12-03T11:32:27Z","timestamp":1733225547000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-78354-8_30"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,4]]},"ISBN":["9783031783531","9783031783548"],"references-count":47,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-78354-8_30","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,4]]},"assertion":[{"value":"4 December 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICPR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kolkata","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 December 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 December 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icpr2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icpr2024.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}