{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,24]],"date-time":"2026-02-24T17:41:25Z","timestamp":1771954885503,"version":"3.50.1"},"publisher-location":"Cham","reference-count":61,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031250552","type":"print"},{"value":"9783031250569","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-25056-9_1","type":"book-chapter","created":{"date-parts":[[2023,2,14]],"date-time":"2023-02-14T12:09:56Z","timestamp":1676376596000},"page":"3-20","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Globally Optimal Event-Based Divergence Estimation for\u00a0Ventral Landing"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0822-3236","authenticated-orcid":false,"given":"Sofia","family":"McLeod","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9311-6392","authenticated-orcid":false,"given":"Gabriele","family":"Meoni","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9846-8423","authenticated-orcid":false,"given":"Dario","family":"Izzo","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9038-085X","authenticated-orcid":false,"given":"Anne","family":"Mergy","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1929-657X","authenticated-orcid":false,"given":"Daqi","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2529-5322","authenticated-orcid":false,"given":"Yasir","family":"Latif","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7790-6423","authenticated-orcid":false,"given":"Ian","family":"Reid","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2423-9342","authenticated-orcid":false,"given":"Tat-Jun","family":"Chin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,15]]},"reference":[{"key":"1_CR1","unstructured":"E-RAFT: dense optical flow from event cameras. github.com\/uzh-rpg\/E-RAFT. Accessed 24 Feb 2022"},{"key":"1_CR2","unstructured":"Event contrast maximization library. https:\/\/github.com\/TimoStoff\/events_contrast_maximization. Accessed 27 Feb 2022"},{"key":"1_CR3","unstructured":"Moon LRO LOLA - Selene Kaguya TC dem merge 60N60S 59m v1. https:\/\/astrogeology.usgs.gov\/search\/map\/Moon\/LRO\/LOLA\/Lunar_LRO_LOLAKaguya_DEMmerge_60N60S_512ppd. Accessed 24 June 2021"},{"key":"1_CR4","unstructured":"Planet and asteroid natural scene generation utility product website. https:\/\/pangu.software\/. Accessed 26 Jan 2022"},{"key":"1_CR5","unstructured":"Vertical landing for micro air vehicles using event-based optical flow dataset. https:\/\/dataverse.nl\/dataset.xhtml?persistentId=hdl:10411\/FBKJFH. Accessed 31 Jan 2022"},{"issue":"7","key":"1_CR6","doi-asserted-by":"publisher","first-page":"1547","DOI":"10.1109\/TPAMI.2020.2986748","volume":"42","author":"M Almatrafi","year":"2020","unstructured":"Almatrafi, M., Baldwin, R., Aizawa, K., Hirakawa, K.: Distance surface for event-based optical flow. IEEE Trans. Pattern Anal. Mach. Intell. 42(7), 1547\u20131556 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1_CR7","doi-asserted-by":"crossref","unstructured":"Bardow, P., Davison, A.J., Leutenegger, S.: Simultaneous optical flow and intensity estimation from an event camera. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 884\u2013892 (2016)","DOI":"10.1109\/CVPR.2016.102"},{"issue":"2","key":"1_CR8","doi-asserted-by":"publisher","first-page":"407","DOI":"10.1109\/TNNLS.2013.2273537","volume":"25","author":"R Benosman","year":"2014","unstructured":"Benosman, R., Clercq, C., Lagorce, X., Ieng, S.H., Bartolozzi, C.: Event-based visual flow. IEEE Trans. Neural Netw. Learn. Syst. 25(2), 407\u2013417 (2014)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"1_CR9","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1016\/j.neunet.2011.11.001","volume":"27","author":"R Benosman","year":"2012","unstructured":"Benosman, R., Ieng, S.H., Clercq, C., Bartolozzi, C., Srinivasan, M.: Asynchronous frameless event-based optical flow. Neural Netw. 27, 32\u201337 (2012)","journal-title":"Neural Netw."},{"issue":"10","key":"1_CR10","doi-asserted-by":"publisher","first-page":"2333","DOI":"10.1109\/JSSC.2014.2342715","volume":"49","author":"C Brandli","year":"2014","unstructured":"Brandli, C., Berner, R., Yang, M., Liu, S.C., Delbruck, T.: A 240 $$\\times $$ 180 130 db 3 $$\\upmu $$s latency global shutter spatiotemporal vision sensor. IEEE J. Solid-State Circuits 49(10), 2333\u20132341 (2014)","journal-title":"IEEE J. Solid-State Circuits"},{"key":"1_CR11","doi-asserted-by":"crossref","unstructured":"Chin, T.J., Bagchi, S., Eriksson, A., van Schaik, A.: Star tracking using an event camera. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (2019)","DOI":"10.1109\/CVPRW.2019.00208"},{"key":"1_CR12","doi-asserted-by":"publisher","first-page":"9","DOI":"10.3389\/fnins.2014.00009","volume":"8","author":"X Clady","year":"2014","unstructured":"Clady, X., et al.: Asynchronous visual event-based time-to-contact. Front. Neurosci. 8, 9 (2014)","journal-title":"Front. Neurosci."},{"key":"1_CR13","doi-asserted-by":"crossref","unstructured":"Dinaux, R., Wessendorp, N., Dupeyroux, J., Croon, G.C.H.E.D.: FAITH: fast iterative half-Plane focus of expansion estimation using optic flow. IEEE Rob. Autom. Lett. 6(4), 7627\u20137634 (2021)","DOI":"10.1109\/LRA.2021.3100153"},{"key":"1_CR14","doi-asserted-by":"crossref","unstructured":"Ding, Z., et al.: Spatio-temporal recurrent networks for event-based optical flow estimation. In: AAAI Conference on Artificial Intelligence, pp. 1\u201313 (2021)","DOI":"10.1609\/aaai.v36i1.19931"},{"key":"1_CR15","doi-asserted-by":"crossref","unstructured":"Falanga, D., Kleber, K., Scaramuzza, D.: Dynamic obstacle avoidance for quadrotors with event cameras. Sci. Rob. 5(40), eaaz9712 (2020)","DOI":"10.1126\/scirobotics.aaz9712"},{"key":"1_CR16","unstructured":"Forsyth, D., Ponce, J.: Computer Vision: A Modern Approach. Prentice hall (2011)"},{"key":"1_CR17","doi-asserted-by":"publisher","unstructured":"Fry, S.N.: Experimental approaches toward a functional understanding of insect flight control. In: Floreano, D., Zufferey, J.C., Srinivasan, M.V., Ellington, C. (eds.) Flying Insects and Robots, pp. 1\u201313. Springer, Heidelberg (2010). https:\/\/doi.org\/10.1007\/978-3-540-89393-6_1","DOI":"10.1007\/978-3-540-89393-6_1"},{"key":"1_CR18","doi-asserted-by":"crossref","unstructured":"Gallego, G., Rebecq, H., Scaramuzza, D.: A unifying contrast maximization framework for event cameras, with applications to motion, depth, and optical flow estimation. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3867\u20133876 (2018)","DOI":"10.1109\/CVPR.2018.00407"},{"key":"1_CR19","doi-asserted-by":"crossref","unstructured":"Gehrig, D., Loquercio, A., Derpanis, K.G., Scaramuzza, D.: End-to-end learning of representations for asynchronous event-based data. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5633\u20135643 (2019)","DOI":"10.1109\/ICCV.2019.00573"},{"key":"1_CR20","doi-asserted-by":"crossref","unstructured":"Gehrig, M., Millh\u00e4usler, M., Gehrig, D., Scaramuzza, D.: E-RAFT: dense optical flow from event cameras. In: 2021 International Conference on 3D Vision (3DV), pp. 197\u2013206 (2021)","DOI":"10.1109\/3DV53792.2021.00030"},{"key":"1_CR21","doi-asserted-by":"crossref","unstructured":"G\u00f3mez Egu\u00edluz, A., Rodr\u00edguez-G\u00f3mez, J.P., Mart\u00ednez-de Dios, J.R., Ollero, A.: Asynchronous event-based line tracking for Time-to-Contact maneuvers in UAS. In: 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5978\u20135985 (2020)","DOI":"10.1109\/IROS45743.2020.9341240"},{"issue":"4","key":"1_CR22","doi-asserted-by":"publisher","first-page":"860","DOI":"10.1109\/TBCAS.2018.2834558","volume":"12","author":"G Haessig","year":"2018","unstructured":"Haessig, G., Cassidy, A., Alvarez, R., Benosman, R., Orchard, G.: Spiking optical flow for event-based sensors using IBM\u2019s TrueNorth neurosynaptic system. IEEE Trans. Biomed. Circuits Syst. 12(4), 860\u2013870 (2018)","journal-title":"IEEE Trans. Biomed. Circuits Syst."},{"key":"1_CR23","unstructured":"Hagenaars, J.J., Paredes-Vall\u00e9s, F., de Croon, G.C.H.E.: Self-supervised learning of event-based optical flow with spiking neural networks. In: Neural Information Processing Systems, October 2021"},{"key":"1_CR24","doi-asserted-by":"publisher","unstructured":"Horst, R., Hoang, T.: Global Optimization: Deterministic Approaches. Springer, Heidelberg (1996). https:\/\/doi.org\/10.1007\/978-3-662-03199-5","DOI":"10.1007\/978-3-662-03199-5"},{"key":"1_CR25","doi-asserted-by":"crossref","unstructured":"Hu, Y., Liu, S.C., Delbruck, T.: V2E: from video frames to realistic DVS events. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (2021)","DOI":"10.1109\/CVPRW53098.2021.00144"},{"key":"1_CR26","unstructured":"Intel: beyond today\u2019s AI. https:\/\/www.intel.com.au\/content\/www\/au\/en\/research\/neuromorphic-computing.html"},{"key":"1_CR27","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"366","DOI":"10.1007\/978-3-030-58526-6_22","volume-title":"Computer Vision \u2013 ECCV 2020","author":"C Lee","year":"2020","unstructured":"Lee, C., Kosta, A.K., Zhu, A.Z., Chaney, K., Daniilidis, K., Roy, K.: Spike-FlowNet: event-based optical flow estimation with energy-efficient hybrid neural networks. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12374, pp. 366\u2013382. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58526-6_22"},{"issue":"2","key":"1_CR28","doi-asserted-by":"publisher","first-page":"566","DOI":"10.1109\/JSSC.2007.914337","volume":"43","author":"P Lichtsteiner","year":"2008","unstructured":"Lichtsteiner, P., Posch, C., Delbruck, T.: A 128 $$\\times $$ 128 120 db 15 $$\\upmu $$s latency asynchronous temporal contrast vision sensor. IEEE J. Solid-State Circuits 43(2), 566\u2013576 (2008)","journal-title":"IEEE J. Solid-State Circuits"},{"key":"1_CR29","doi-asserted-by":"crossref","unstructured":"Liu, D., Parra, A., Chin, T.J.: Globally optimal contrast maximisation for event-based motion estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6349\u20136358 (2020)","DOI":"10.1109\/CVPR42600.2020.00638"},{"key":"1_CR30","doi-asserted-by":"crossref","unstructured":"Liu, M., Delbruck, T.: Adaptive time-slice block-matching optical flow algorithm for dynamic vision sensors. In: British Machine Vision Conference (BMVC) (2018)","DOI":"10.1109\/ISCAS.2017.8050295"},{"key":"1_CR31","unstructured":"Lucas, B.D., Kanade, T.: An iterative image registration technique with an application to stereo vision. In: International Joint Conference on Artificial Intelligence, pp. 674\u2013679 (1981)"},{"key":"1_CR32","doi-asserted-by":"crossref","unstructured":"Mueggler, E., Huber, B., Scaramuzza, D.: Event-based, 6-DOF pose tracking for high-speed maneuvers. In: 2014 IEEE\/RSJ International Conference on Intelligent Robots and Systems, pp. 2761\u20132768 (2014)","DOI":"10.1109\/IROS.2014.6942940"},{"key":"1_CR33","unstructured":"NASA\/JPL-Caltech\/University of Arizona: Decoding a geological message (2017). https:\/\/www.nasa.gov\/sites\/default\/files\/thumbnails\/image\/pia21759.jpg"},{"key":"1_CR34","unstructured":"NASA\/JPL-Caltech\/University of Arizona: Big fans (2018). https:\/\/www.nasa.gov\/image-feature\/jpl\/pia22332\/big-fans"},{"key":"1_CR35","doi-asserted-by":"crossref","unstructured":"Orchard, G., Bartolozzi, C., Indiveri, G.: Applying neuromorphic vision sensors to planetary landing tasks. In: IEEE Biomedical Circuits and Systems Conference, pp. 201\u2013204 (2009)","DOI":"10.1109\/BIOCAS.2009.5372048"},{"key":"1_CR36","doi-asserted-by":"crossref","unstructured":"Orchard, G., Benosman, R., Etienne-Cummings, R., Thakor, N.V.: A spiking neural network architecture for visual motion estimation. In: IEEE Biomedical Circuits and Systems Conference (BioCAS), pp. 298\u2013301 (2013)","DOI":"10.1109\/BioCAS.2013.6679698"},{"key":"1_CR37","doi-asserted-by":"crossref","unstructured":"Pan, L., Liu, M., Hartley, R.: Single image optical flow estimation with an event camera. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1669\u20131678 (2020)","DOI":"10.1109\/CVPR42600.2020.00174"},{"key":"1_CR38","doi-asserted-by":"crossref","unstructured":"Paredes-Vall\u00e9s, F., de Croon, G.C.H.E.: Back to event basics: self-supervised learning of image reconstruction for event cameras via photometric constancy. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2021)","DOI":"10.1109\/CVPR46437.2021.00345"},{"issue":"8","key":"1_CR39","doi-asserted-by":"publisher","first-page":"2051","DOI":"10.1109\/TPAMI.2019.2903179","volume":"42","author":"F Paredes-Vall\u00e9s","year":"2020","unstructured":"Paredes-Vall\u00e9s, F., Scheper, K.Y.W., de Croon, G.C.H.E.: Unsupervised learning of a hierarchical spiking neural network for optical flow estimation: from events to global motion perception. IEEE Trans. Pattern Anal. Mach. Intell. 42(8), 2051\u20132064 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1_CR40","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"51","DOI":"10.1007\/978-3-030-58574-7_4","volume-title":"Computer Vision \u2013 ECCV 2020","author":"X Peng","year":"2020","unstructured":"Peng, X., Wang, Y., Gao, L., Kneip, L.: Globally-optimal event camera motion estimation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12371, pp. 51\u201367. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58574-7_4"},{"key":"1_CR41","doi-asserted-by":"crossref","unstructured":"Pijnacker Hordijk, B.J., Scheper, K.Y.W., de Croon, G.C.H.E.: Vertical landing for micro air vehicles using event-based optical flow. J. Field Rob. 35(1), 69\u201390 (2018)","DOI":"10.1002\/rob.21764"},{"issue":"1","key":"1_CR42","doi-asserted-by":"publisher","first-page":"259","DOI":"10.1109\/JSSC.2010.2085952","volume":"46","author":"C Posch","year":"2011","unstructured":"Posch, C., Matolin, D., Wohlgenannt, R.: A QVGA 143 db dynamic range frame-free PWM image sensor with lossless Pixel-Level video compression and time-domain CDS. IEEE J. Solid-State Circuits 46(1), 259\u2013275 (2011)","journal-title":"IEEE J. Solid-State Circuits"},{"key":"1_CR43","doi-asserted-by":"crossref","unstructured":"Rebecq, H., Ranftl, R., Koltun, V., Scaramuzza, D.: Events-to-video: bringing modern computer vision to event cameras. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.00398"},{"issue":"6","key":"1_CR44","doi-asserted-by":"publisher","first-page":"1964","DOI":"10.1109\/TPAMI.2019.2963386","volume":"43","author":"H Rebecq","year":"2019","unstructured":"Rebecq, H., Ranftl, R., Koltun, V., Scaramuzza, D.: High speed and high dynamic range video with an event camera. IEEE Trans. Pattern Anal. Mach. Intell. 43(6), 1964\u20131980 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1_CR45","doi-asserted-by":"crossref","unstructured":"Sanket, N.J., et al.: EVDodgeNet: deep dynamic obstacle dodging with event cameras. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 10651\u201310657 (2020)","DOI":"10.1109\/ICRA40945.2020.9196877"},{"key":"1_CR46","doi-asserted-by":"crossref","unstructured":"Scheerlinck, C., Rebecq, H., Gehrig, D., Barnes, N., Mahony, R., Scaramuzza, D.: Fast image reconstruction with an event camera. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 156\u2013163 (2020)","DOI":"10.1109\/WACV45572.2020.9093366"},{"key":"1_CR47","doi-asserted-by":"crossref","unstructured":"Sikorski, O., Izzo, D., Meoni, G.: Event-based spacecraft landing using time-to-contact. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1941\u20131950 (2021)","DOI":"10.1109\/CVPRW53098.2021.00222"},{"issue":"Pt 1","key":"1_CR48","doi-asserted-by":"publisher","first-page":"237","DOI":"10.1242\/jeb.199.1.237","volume":"199","author":"M Srinivasan","year":"1996","unstructured":"Srinivasan, M., Zhang, S., Lehrer, M., Collett, T.: Honeybee navigation EN route to the goal: visual flight control and odometry. J. Exp. Biol. 199(Pt 1), 237\u2013244 (1996)","journal-title":"J. Exp. Biol."},{"key":"1_CR49","first-page":"413","volume":"91","author":"MV Srinivasan","year":"2011","unstructured":"Srinivasan, M.V.: Honeybees as a model for the study of visually guided flight, navigation, and biologically inspired robotics. Phys. Rev. 91, 413\u2013406 (2011)","journal-title":"Phys. Rev."},{"key":"1_CR50","doi-asserted-by":"publisher","unstructured":"Srinivasan, M.V., Thurrowgood, S., Soccol, D.: From visual guidance in flying insects to autonomous aerial vehicles. In: Floreano, D., Zufferey, J.C., Srinivasan, M.V., Ellington, C. (eds.) Flying Insects and Robots, pp. 15\u201328. Springer, Heidelberg (2010). https:\/\/doi.org\/10.1007\/978-3-540-89393-6_2","DOI":"10.1007\/978-3-540-89393-6_2"},{"key":"1_CR51","unstructured":"Stoffregen, T., Kleeman, L.: Simultaneous optical flow and segmentation (SOFAS) using dynamic vision sensor. In: Australasian Conference on Robotics and Automation (2018)"},{"key":"1_CR52","doi-asserted-by":"crossref","unstructured":"Stoffregen, T., Gallego, G., Drummond, T., Kleeman, L., Scaramuzza, D.: Event-based motion segmentation by motion compensation. In: International Conference on Computer Vision, pp. 7243\u20137252 (2019)","DOI":"10.1109\/ICCV.2019.00734"},{"key":"1_CR53","doi-asserted-by":"crossref","unstructured":"Stoffregen, T., Kleeman, L.: Event cameras, contrast maximization and reward functions: an analysis. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12300\u201312308 (2019)","DOI":"10.1109\/CVPR.2019.01258"},{"issue":"Pt 3","key":"1_CR54","doi-asserted-by":"publisher","first-page":"327","DOI":"10.1242\/jeb.205.3.327","volume":"205","author":"LF Tammero","year":"2002","unstructured":"Tammero, L.F., Dickinson, M.H.: The influence of visual landscape on the free flight behavior of the fruit fly drosophila melanogaster. J. Exp. Biol. 205(Pt 3), 327\u2013343 (2002)","journal-title":"J. Exp. Biol."},{"key":"1_CR55","doi-asserted-by":"crossref","unstructured":"Valette, F., Ruffier, F., Viollet, S., Seidl, T.: Biomimetic optic flow sensing applied to a lunar landing scenario. In: 2010 IEEE International Conference on Robotics and Automation, pp. 2253\u20132260 (2010)","DOI":"10.1109\/ROBOT.2010.5509364"},{"key":"1_CR56","unstructured":"Medici, V., Orchard, G., Ammann, S., Indiveri, G., Fry, S.N.: Neuromorphic computation of optic flow data. Technical report, European Space Agency, Advanced Concepts Team (2010)"},{"issue":"2","key":"1_CR57","doi-asserted-by":"publisher","first-page":"994","DOI":"10.1109\/LRA.2018.2793357","volume":"3","author":"AR Vidal","year":"2018","unstructured":"Vidal, A.R., Rebecq, H., Horstschaefer, T., Scaramuzza, D.: Ultimate SLAM? Combining events, images, and IMU for robust visual SLAM in HDR and high-speed scenarios. IEEE Rob. Autom. Lett. 3(2), 994\u20131001 (2018)","journal-title":"IEEE Rob. Autom. Lett."},{"key":"1_CR58","doi-asserted-by":"crossref","unstructured":"Ye, C., Mitrokhin, A., Ferm\u00fcller, C., Yorke, J.A., Aloimonos, Y.: Unsupervised learning of dense optical flow, depth and egomotion with Event-Based sensors. In: 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5831\u20135838 (2020)","DOI":"10.1109\/IROS45743.2020.9341224"},{"key":"1_CR59","unstructured":"Zhou, Y., Gallego, G., Lu, X., Liu, S., Shen, S.: Event-Based motion segmentation with Spatio-Temporal graph cuts. IEEE Trans. Neural Netw. Learn. Syst. (2020)"},{"key":"1_CR60","doi-asserted-by":"crossref","unstructured":"Zhu, A.Z., Yuan, L., Chaney, K., Daniilidis, K.: EV-FlowNet: self-Supervised optical flow estimation for event-based cameras. Rob. Sci. Syst. (2018)","DOI":"10.15607\/RSS.2018.XIV.062"},{"key":"1_CR61","doi-asserted-by":"crossref","unstructured":"Zhu, A.Z., Yuan, L., Chaney, K., Daniilidis, K.: Unsupervised event-based learning of optical flow, depth, and egomotion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 989\u2013997 (2019). openaccess.thecvf.com","DOI":"10.1109\/CVPR.2019.00108"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-25056-9_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T18:27:49Z","timestamp":1710268069000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-25056-9_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031250552","9783031250569"],"references-count":61,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-25056-9_1","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"15 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}