{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,30]],"date-time":"2025-12-30T08:51:50Z","timestamp":1767084710280,"version":"3.40.3"},"publisher-location":"Cham","reference-count":73,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031200670"},{"type":"electronic","value":"9783031200687"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20068-7_27","type":"book-chapter","created":{"date-parts":[[2022,11,10]],"date-time":"2022-11-10T08:06:38Z","timestamp":1668067598000},"page":"470-488","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Boosting Event Stream Super-Resolution with\u00a0a\u00a0Recurrent Neural Network"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1042-8903","authenticated-orcid":false,"given":"Wenming","family":"Weng","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0788-8826","authenticated-orcid":false,"given":"Yueyi","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9787-7460","authenticated-orcid":false,"given":"Zhiwei","family":"Xiong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,11]]},"reference":[{"key":"27_CR1","doi-asserted-by":"crossref","unstructured":"Amir, A., et al.: A low power, fully event-based gesture recognition system. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.781"},{"key":"27_CR2","doi-asserted-by":"crossref","unstructured":"Bevilacqua, M., Roumy, A., Guillemot, C., Alberi-Morel, M.L.: Low-complexity single-image super-resolution based on nonnegative neighbor embedding (2012)","DOI":"10.5244\/C.26.135"},{"key":"27_CR3","doi-asserted-by":"crossref","unstructured":"Bi, Y., Chadha, A., Abbas, A., Bourtsoulatze, E., Andreopoulos, Y.: Graph-based object classification for neuromorphic vision sensing. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00058"},{"issue":"10","key":"27_CR4","doi-asserted-by":"publisher","first-page":"2333","DOI":"10.1109\/JSSC.2014.2342715","volume":"49","author":"C Brandli","year":"2014","unstructured":"Brandli, C., Berner, R., Yang, M., Liu, S.C., Delbruck, T.: A 240$$\\times $$ 180 130 db 3 $$\\mu $$s latency global shutter spatiotemporal vision sensor. IEEE J. Solid-State Circuits 49(10), 2333\u20132341 (2014)","journal-title":"IEEE J. Solid-State Circuits"},{"key":"27_CR5","doi-asserted-by":"crossref","unstructured":"Caballero, J., et al.: Real-time video super-resolution with spatio-temporal networks and motion compensation. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.304"},{"key":"27_CR6","doi-asserted-by":"crossref","unstructured":"Chan, K.C., Wang, X., Yu, K., Dong, C., Loy, C.C.: Basicvsr: the search for essential components in video super-resolution and beyond. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00491"},{"key":"27_CR7","unstructured":"Chang, H., Yeung, D.Y., Xiong, Y.: Super-resolution through neighbor embedding. In: CVPR (2004)"},{"key":"27_CR8","doi-asserted-by":"crossref","unstructured":"Cho, K., et al.: Learning phrase representations using rnn encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078 (2014)","DOI":"10.3115\/v1\/D14-1179"},{"key":"27_CR9","unstructured":"Choi, J., Yoon, K.J., et al.: Learning to super resolve intensity images from events. In: CVPR (2020)"},{"key":"27_CR10","doi-asserted-by":"crossref","unstructured":"Dai, J., et al.: Deformable convolutional networks. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.89"},{"key":"27_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"184","DOI":"10.1007\/978-3-319-10593-2_13","volume-title":"Computer Vision \u2013 ECCV 2014","author":"C Dong","year":"2014","unstructured":"Dong, C., Loy, C.C., He, K., Tang, X.: Learning a deep convolutional network for image super-resolution. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8692, pp. 184\u2013199. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10593-2_13"},{"key":"27_CR12","doi-asserted-by":"crossref","unstructured":"Duan, P., Wang, Z.W., Zhou, X., Ma, Y., Shi, B.: Eventzoom: learning to denoise and super resolve neuromorphic events. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01263"},{"key":"27_CR13","doi-asserted-by":"publisher","unstructured":"Gallego, G., et al.: Event-based vision: a survey. IEEE Trans. Pattern Anal. Mach. Intell. (2020). https:\/\/doi.org\/10.1109\/TPAMI.2020.3008413","DOI":"10.1109\/TPAMI.2020.3008413"},{"issue":"10","key":"27_CR14","doi-asserted-by":"publisher","first-page":"2402","DOI":"10.1109\/TPAMI.2017.2769655","volume":"40","author":"G Gallego","year":"2017","unstructured":"Gallego, G., Lund, J.E., Mueggler, E., Rebecq, H., Delbruck, T., Scaramuzza, D.: Event-based, 6-DOF camera tracking from photometric depth maps. IEEE Trans. Pattern Anal. Mach. Intell. 40(10), 2402\u20132412 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"27_CR15","doi-asserted-by":"crossref","unstructured":"Gallego, G., Rebecq, H., Scaramuzza, D.: A unifying contrast maximization framework for event cameras, with applications to motion, depth, and optical flow estimation. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00407"},{"key":"27_CR16","doi-asserted-by":"crossref","unstructured":"Gehrig, D., Gehrig, M., Hidalgo-Carri\u00f3, J., Scaramuzza, D.: Video to events: recycling video datasets for event cameras. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00364"},{"key":"27_CR17","doi-asserted-by":"crossref","unstructured":"Gehrig, D., Loquercio, A., Derpanis, K.G., Scaramuzza, D.: End-to-end learning of representations for asynchronous event-based data. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00573"},{"key":"27_CR18","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511815706","volume-title":"Spiking Neuron Models: Single Neurons, Populations, Plasticity","author":"W Gerstner","year":"2002","unstructured":"Gerstner, W., Kistler, W.M.: Spiking Neuron Models: Single Neurons, Populations, Plasticity. Cambridge University Press, Cambridge (2002)"},{"key":"27_CR19","doi-asserted-by":"crossref","unstructured":"Glasner, D., Bagon, S., Irani, M.: Super-resolution from a single image. In: ICCV (2009)","DOI":"10.1109\/ICCV.2009.5459271"},{"key":"27_CR20","doi-asserted-by":"crossref","unstructured":"Gu, C., Learned-Miller, E., Sheldon, D., Gallego, G., Bideau, P.: The spatio-temporal poisson point process: a simple model for the alignment of event camera data. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01324"},{"key":"27_CR21","doi-asserted-by":"crossref","unstructured":"Haris, M., Shakhnarovich, G., Ukita, N.: Recurrent back-projection network for video super-resolution. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00402"},{"key":"27_CR22","doi-asserted-by":"crossref","unstructured":"He, W., et al.: Timereplayer: unlocking the potential of event cameras for video interpolation. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01728"},{"issue":"18","key":"27_CR23","doi-asserted-by":"publisher","first-page":"23366","DOI":"10.1364\/OE.26.023366","volume":"26","author":"S Heist","year":"2018","unstructured":"Heist, S., Zhang, C., Reichwald, K., K\u00fchmstedt, P., Notni, G., T\u00fcnnermann, A.: 5D hyperspectral imaging: fast and accurate measurement of surface shape and spectral characteristics using structured light. Opt. Express 26(18), 23366\u201323379 (2018)","journal-title":"Opt. Express"},{"key":"27_CR24","doi-asserted-by":"crossref","unstructured":"Huang, J.B., Singh, A., Ahuja, N.: Single image super-resolution from transformed self-exemplars. In: CVPR (2015)","DOI":"10.1109\/CVPR.2015.7299156"},{"key":"27_CR25","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"645","DOI":"10.1007\/978-3-030-58610-2_38","volume-title":"Computer Vision \u2013 ECCV 2020","author":"T Isobe","year":"2020","unstructured":"Isobe, T., Jia, X., Gu, S., Li, S., Wang, S., Tian, Q.: Video super-resolution with recurrent structure-detail network. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12357, pp. 645\u2013660. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58610-2_38"},{"key":"27_CR26","unstructured":"Isobe, T., Zhu, F., Jia, X., Wang, S.: Revisiting temporal modeling for video super-resolution. In: BMVC (2020)"},{"key":"27_CR27","doi-asserted-by":"crossref","unstructured":"Jiang, Z., Zhang, Y., Zou, D., Ren, J., Lv, J., Liu, Y.: Learning event-based motion deblurring. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00338"},{"key":"27_CR28","doi-asserted-by":"crossref","unstructured":"Jo, Y., Oh, S.W., Kang, J., Kim, S.J.: Deep video super-resolution network using dynamic upsampling filters without explicit motion compensation. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00340"},{"key":"27_CR29","doi-asserted-by":"crossref","unstructured":"Kiani Galoogahi, H., Fagg, A., Huang, C., Ramanan, D., Lucey, S.: Need for speed: a benchmark for higher frame rate object tracking. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.128"},{"key":"27_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"349","DOI":"10.1007\/978-3-319-46466-4_21","volume-title":"Computer Vision \u2013 ECCV 2016","author":"H Kim","year":"2016","unstructured":"Kim, H., Leutenegger, S., Davison, A.J.: Real-time 3D reconstruction and 6-DoF tracking with an event camera. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9910, pp. 349\u2013364. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46466-4_21"},{"issue":"4","key":"27_CR31","first-page":"1","volume":"31","author":"MH Kim","year":"2012","unstructured":"Kim, M.H., et al.: 3D imaging spectroscopy for measuring hyperspectral patterns on solid objects. ACM Trans. Graph. (TOG) 31(4), 1\u201311 (2012)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"27_CR32","doi-asserted-by":"crossref","unstructured":"Ledig, C., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.19"},{"key":"27_CR33","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1016\/j.neucom.2018.12.048","volume":"335","author":"H Li","year":"2019","unstructured":"Li, H., Li, G., Shi, L.: Super-resolution of spatiotemporal event-stream image. Neurocomputing 335, 206\u2013214 (2019)","journal-title":"Neurocomputing"},{"key":"27_CR34","doi-asserted-by":"crossref","unstructured":"Li, S., Feng, Y., Li, Y., Jiang, Y., Zou, C., Gao, Y.: Event stream super-resolution via spatiotemporal constraint learning. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00444"},{"key":"27_CR35","doi-asserted-by":"crossref","unstructured":"Li, Z., Yang, J., Liu, Z., Yang, X., Jeon, G., Wu, W.: Feedback network for image super-resolution. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00399"},{"key":"27_CR36","doi-asserted-by":"crossref","unstructured":"Lim, B., Son, S., Kim, H., Nah, S., Mu Lee, K.: Enhanced deep residual networks for single image super-resolution. In: CVPRW (2017)","DOI":"10.1109\/CVPRW.2017.151"},{"key":"27_CR37","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"695","DOI":"10.1007\/978-3-030-58598-3_41","volume-title":"Computer Vision \u2013 ECCV 2020","author":"S Lin","year":"2020","unstructured":"Lin, S., et al.: Learning event-driven video deblurring and interpolation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12353, pp. 695\u2013710. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58598-3_41"},{"key":"27_CR38","doi-asserted-by":"crossref","unstructured":"Liu, D., Parra, A., Chin, T.J.: Globally optimal contrast maximisation for event-based motion estimation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00638"},{"key":"27_CR39","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"415","DOI":"10.1007\/978-3-030-58598-3_25","volume-title":"Computer Vision \u2013 ECCV 2020","author":"N Messikommer","year":"2020","unstructured":"Messikommer, N., Gehrig, D., Loquercio, A., Scaramuzza, D.: Event-based asynchronous sparse convolutional networks. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12353, pp. 415\u2013431. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58598-3_25"},{"issue":"10","key":"27_CR40","doi-asserted-by":"publisher","first-page":"2028","DOI":"10.1109\/TPAMI.2015.2392947","volume":"37","author":"G Orchard","year":"2015","unstructured":"Orchard, G., Meyer, C., Etienne-Cummings, R., Posch, C., Thakor, N., Benosman, R.: Hfirst: a temporal approach to object recognition. IEEE Trans. Pattern Anal. Mach. Intell. 37(10), 2028\u20132040 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"27_CR41","doi-asserted-by":"crossref","unstructured":"Pan, L., Scheerlinck, C., Yu, X., Hartley, R., Liu, M., Dai, Y.: Bringing a blurry frame alive at high frame-rate with an event camera. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00698"},{"issue":"8","key":"27_CR42","doi-asserted-by":"publisher","first-page":"2051","DOI":"10.1109\/TPAMI.2019.2903179","volume":"42","author":"F Paredes-Vall\u00e9s","year":"2019","unstructured":"Paredes-Vall\u00e9s, F., Scheper, K.Y., de Croon, G.C.: Unsupervised learning of a hierarchical spiking neural network for optical flow estimation: from events to global motion perception. IEEE Trans. Pattern Anal. Mach. Intell. 42(8), 2051\u20132064 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"27_CR43","doi-asserted-by":"publisher","first-page":"566","DOI":"10.1109\/JSSC.2007.914337","volume":"43","author":"L Patrick","year":"2008","unstructured":"Patrick, L., Posch, C., Delbruck, T.: A 128x 128 120 db 15$$\\mu $$s latency asynchronous temporal contrast vision sensor. IEEE J. Solid-State Circuits 43, 566\u2013576 (2008)","journal-title":"IEEE J. Solid-State Circuits"},{"key":"27_CR44","doi-asserted-by":"crossref","unstructured":"Rebecq, H., Gallego, G., Scaramuzza, D.: EMVS: event-based multi-view stereo. In: BMVC (2016)","DOI":"10.5244\/C.30.63"},{"issue":"6","key":"27_CR45","doi-asserted-by":"publisher","first-page":"1964","DOI":"10.1109\/TPAMI.2019.2963386","volume":"43","author":"H Rebecq","year":"2019","unstructured":"Rebecq, H., Ranftl, R., Koltun, V., Scaramuzza, D.: High speed and high dynamic range video with an event camera. IEEE Trans. Pattern Anal. Mach. Intell. 43(6), 1964\u20131980 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"27_CR46","doi-asserted-by":"crossref","unstructured":"Schaefer, S., Gehrig, D., Scaramuzza, D.: AEGNN: asynchronous event-based graph neural networks. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01205"},{"key":"27_CR47","doi-asserted-by":"crossref","unstructured":"Sironi, A., Brambilla, M., Bourdis, N., Lagorce, X., Benosman, R.: HATS: histograms of averaged time surfaces for robust event-based object classification. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00186"},{"key":"27_CR48","doi-asserted-by":"crossref","unstructured":"Tian, Y., Zhang, Y., Fu, Y., Xu, C.: TDAN: temporally-deformable alignment network for video super-resolution. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00342"},{"key":"27_CR49","doi-asserted-by":"crossref","unstructured":"Timofte, R., De Smet, V., Van Gool, L.: Anchored neighborhood regression for fast example-based super-resolution. In: ICCV (2013)","DOI":"10.1109\/ICCV.2013.241"},{"key":"27_CR50","doi-asserted-by":"crossref","unstructured":"Timofte, R., De Smet, V., Van Gool, L.: A+: adjusted anchored neighborhood regression for fast super-resolution. In: ACCV (2014)","DOI":"10.1109\/ICCV.2013.241"},{"key":"27_CR51","doi-asserted-by":"crossref","unstructured":"Tong, T., Li, G., Liu, X., Gao, Q.: Image super-resolution using dense skip connections. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.514"},{"key":"27_CR52","doi-asserted-by":"crossref","unstructured":"Tulyakov, S., Bochicchio, A., Gehrig, D., Georgoulis, S., Li, Y., Scaramuzza, D.: Time lens++: event-based frame interpolation with parametric non-linear flow and multi-scale fusion. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01723"},{"key":"27_CR53","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"155","DOI":"10.1007\/978-3-030-58601-0_10","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Wang","year":"2020","unstructured":"Wang, B., He, J., Yu, L., Xia, G.-S., Yang, W.: Event enhanced high-quality image recovery. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12358, pp. 155\u2013171. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58601-0_10"},{"key":"27_CR54","doi-asserted-by":"crossref","unstructured":"Wang, L., Ho, Y.S., Yoon, K.J., et al.: Event-based high dynamic range image and very high frame rate video generation using conditional generative adversarial networks. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01032"},{"key":"27_CR55","doi-asserted-by":"crossref","unstructured":"Wang, L., Kim, T.K., Yoon, K.J.: Eventsr: from asynchronous events to image reconstruction, restoration, and super-resolution via end-to-end adversarial learning. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00834"},{"issue":"4","key":"27_CR56","first-page":"1","volume":"36","author":"TC Wang","year":"2017","unstructured":"Wang, T.C., Zhu, J.Y., Kalantari, N.K., Efros, A.A., Ramamoorthi, R.: Light field video capture using a learning-based hybrid imaging system. ACM Trans. Graph. (TOG) 36(4), 1\u201313 (2017)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"27_CR57","doi-asserted-by":"crossref","unstructured":"Wang, X., Chan, K.C., Yu, K., Dong, C., Change Loy, C.: EDVR: video restoration with enhanced deformable convolutional networks. In: CVPRW (2019)","DOI":"10.1109\/CVPRW.2019.00247"},{"key":"27_CR58","doi-asserted-by":"crossref","unstructured":"Wang, Z.W., Duan, P., Cossairt, O., Katsaggelos, A., Huang, T., Shi, B.: Joint filtering of intensity images and neuromorphic events for high-resolution noise-robust imaging. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00168"},{"key":"27_CR59","doi-asserted-by":"crossref","unstructured":"Weng, W., Zhang, Y., Xiong, Z.: Event-based video reconstruction using transformer. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00256"},{"key":"27_CR60","doi-asserted-by":"crossref","unstructured":"Xiang, X., Tian, Y., Zhang, Y., Fu, Y., Allebach, J.P., Xu, C.: Zooming slow-mo: fast and accurate one-stage space-time video super-resolution. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00343"},{"key":"27_CR61","doi-asserted-by":"crossref","unstructured":"Xiao, Z., Fu, X., Huang, J., Cheng, Z., Xiong, Z.: Space-time distillation for video super-resolution. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00215"},{"key":"27_CR62","doi-asserted-by":"crossref","unstructured":"Xiao, Z., Xiong, Z., Fu, X., Liu, D., Zha, Z.J.: Space-time video super-resolution using temporal profiles. In: ACM MM (2020)","DOI":"10.1145\/3394171.3413667"},{"key":"27_CR63","doi-asserted-by":"crossref","unstructured":"Xu, G., Xu, J., Li, Z., Wang, L., Sun, X., Cheng, M.M.: Temporal modulation network for controllable space-time video super-resolution. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00632"},{"key":"27_CR64","unstructured":"Yang, J., Wright, J., Huang, T., Ma, Y.: Image super-resolution as sparse representation of raw image patches. In: CVPR (2008)"},{"issue":"11","key":"27_CR65","doi-asserted-by":"publisher","first-page":"2861","DOI":"10.1109\/TIP.2010.2050625","volume":"19","author":"J Yang","year":"2010","unstructured":"Yang, J., Wright, J., Huang, T.S., Ma, Y.: Image super-resolution via sparse representation. IEEE Trans. Image Process. 19(11), 2861\u20132873 (2010)","journal-title":"IEEE Trans. Image Process."},{"issue":"26","key":"27_CR66","doi-asserted-by":"publisher","first-page":"38312","DOI":"10.1364\/OE.27.038312","volume":"27","author":"M Yao","year":"2019","unstructured":"Yao, M., Xiong, Z., Wang, L., Liu, D., Chen, X.: Spectral-depth imaging with deep learning based reconstruction. Opt. Express 27(26), 38312\u201338325 (2019)","journal-title":"Opt. Express"},{"key":"27_CR67","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"711","DOI":"10.1007\/978-3-642-27413-8_47","volume-title":"Curves and Surfaces","author":"R Zeyde","year":"2012","unstructured":"Zeyde, R., Elad, M., Protter, M.: On single image scale-up using sparse-representations. In: Boissonnat, J.-D., et al. (eds.) Curves and Surfaces 2010. LNCS, vol. 6920, pp. 711\u2013730. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-27413-8_47"},{"key":"27_CR68","doi-asserted-by":"crossref","unstructured":"Zhang, X., Liao, W., Yu, L., Yang, W., Xia, G.S.: Event-based synthetic aperture imaging with a hybrid network. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01401"},{"key":"27_CR69","doi-asserted-by":"crossref","unstructured":"Zhang, X., Yu, L.: Unifying motion deblurring and frame interpolation with events. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01724"},{"key":"27_CR70","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Gallego, G., Rebecq, H., Kneip, L., Li, H., Scaramuzza, D.: Semi-dense 3D reconstruction with a stereo event camera. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01246-5_15"},{"key":"27_CR71","doi-asserted-by":"crossref","unstructured":"Zhu, A.Z., Yuan, L., Chaney, K., Daniilidis, K.: Unsupervised event-based learning of optical flow, depth, and egomotion. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00108"},{"key":"27_CR72","doi-asserted-by":"crossref","unstructured":"Zhu, X., Hu, H., Lin, S., Dai, J.: Deformable convnets V2: more deformable, better results. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00953"},{"key":"27_CR73","doi-asserted-by":"crossref","unstructured":"Zihao Zhu, A., Atanasov, N., Daniilidis, K.: Event-based visual inertial odometry. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.616"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20068-7_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,10]],"date-time":"2022-11-10T08:18:45Z","timestamp":1668068325000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20068-7_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031200670","9783031200687"],"references-count":73,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20068-7_27","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"11 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}