{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T06:03:26Z","timestamp":1774677806473,"version":"3.50.1"},"reference-count":60,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,8,30]],"date-time":"2024-08-30T00:00:00Z","timestamp":1724976000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,8,30]],"date-time":"2024-08-30T00:00:00Z","timestamp":1724976000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100002855","name":"Ministry of Science and Technology of the People\u2019s Republic of China","doi-asserted-by":"publisher","award":["2020AAA0108202"],"award-info":[{"award-number":["2020AAA0108202"]}],"id":[{"id":"10.13039\/501100002855","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61931012"],"award-info":[{"award-number":["61931012"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62088102"],"award-info":[{"award-number":["62088102"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2025,3]]},"DOI":"10.1007\/s11263-024-02198-1","type":"journal-article","created":{"date-parts":[[2024,8,30]],"date-time":"2024-08-30T11:03:18Z","timestamp":1725015798000},"page":"991-1011","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Lightweight High-Speed Photography Built on Coded Exposure and Implicit Neural Representation of Videos"],"prefix":"10.1007","volume":"133","author":[{"given":"Zhihong","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Runzhao","family":"Yang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3426-1634","authenticated-orcid":false,"given":"Jinli","family":"Suo","sequence":"additional","affiliation":[]},{"given":"Yuxiao","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Qionghai","family":"Dai","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,30]]},"reference":[{"key":"2198_CR1","doi-asserted-by":"crossref","unstructured":"Agrawal, A., & Raskar, R. (2009). Optimal single image capture for motion deblurring. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, pp 2560\u20132567","DOI":"10.1109\/CVPR.2009.5206546"},{"key":"2198_CR2","doi-asserted-by":"crossref","unstructured":"Agrawal, A., & Xu, Y. (2009). Coded exposure deblurring: Optimized codes for PSF estimation and invertibility. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, pp 2066\u20132073","DOI":"10.1109\/CVPR.2009.5206685"},{"key":"2198_CR3","doi-asserted-by":"crossref","unstructured":"Agrawal, A., Xu, Y., & Raskar, R. (2009). Invertible motion blur in video. In: ACM SIGGRAPH 2009 papers, ACM, pp 1\u20138","DOI":"10.1145\/1576246.1531401"},{"key":"2198_CR4","doi-asserted-by":"crossref","unstructured":"Argaw, D. M., Kim, J., Rameau, F., Zhang, C., & Kweon, I. S. (2021). Restoration of video frames from a single blurred image with motion understanding. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp 701\u2013710","DOI":"10.1109\/CVPRW53098.2021.00079"},{"key":"2198_CR5","doi-asserted-by":"crossref","unstructured":"Charbonnier, P., Blanc-Feraud, L., Aubert, G., & Barlaud, M. (1994). Two deterministic half-quadratic regularization algorithms for computed imaging. In: 1994 IEEE International Conference on Image Processing (ICIP), IEEE Comput. Soc. Press, vol\u00a02, pp 168\u2013172","DOI":"10.1109\/ICIP.1994.413553"},{"key":"2198_CR6","doi-asserted-by":"crossref","unstructured":"Chen, H., Gu, J., Gallo, O., Liu, M. Y., Veeraraghavan, A., & Kautz, J. (2018). Reblur2Deblur: Deblurring videos via self-supervised learning. In: 2018 IEEE International Conference on Computational Photography (ICCP), IEEE, pp 1\u20139","DOI":"10.1109\/ICCPHOT.2018.8368468"},{"key":"2198_CR7","first-page":"21557","volume":"34","author":"H Chen","year":"2021","unstructured":"Chen, H., He, B., Wang, H., Ren, Y., Lim, S. N., & Shrivastava, A. (2021). NeRV: Neural representations for videos. Advances in Neural Information Processing Systems, 34, 21557\u201321568.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2198_CR8","doi-asserted-by":"crossref","unstructured":"Chen, H., Gwilliam, M., Lim. S. N., & Shrivastava, A. (2023). HNeRV: A hybrid neural representation for videos. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR).","DOI":"10.1109\/CVPR52729.2023.00990"},{"key":"2198_CR9","doi-asserted-by":"crossref","unstructured":"Chen, Z., Chen, Y., Liu, J., Xu, X., Goel, V., Wang, Z., Shi, H., & Wang, X. (2022). VideoINR: Learning video implicit neural representation for continuous space-time super-resolution. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 2047\u20132057","DOI":"10.1109\/CVPR52688.2022.00209"},{"key":"2198_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.optlaseng.2020.106489","volume":"139","author":"G Cui","year":"2021","unstructured":"Cui, G., Ye, X., Zhao, J., Zhu, L., Chen, Y., & Zhang, Y. (2021). An effective coded exposure photography framework using optimal fluttering pattern generation. Optics and Lasers in Engineering, 139, 106489.","journal-title":"Optics and Lasers in Engineering"},{"issue":"4","key":"2198_CR11","doi-asserted-by":"publisher","first-page":"1380","DOI":"10.1109\/TPAMI.2019.2946567","volume":"43","author":"C Deng","year":"2021","unstructured":"Deng, C., Zhang, Y., Mao, Y., Fan, J., Suo, J., Zhang, Z., & Dai, Q. (2021). Sinusoidal sampling enhanced compressive camera for high speed imaging. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(4), 1380\u20131393.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"issue":"2s","key":"2198_CR12","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3556544","volume":"19","author":"J Dong","year":"2023","unstructured":"Dong, J., Ota, K., & Dong, M. (2023). Video frame interpolation: A comprehensive survey. ACM Transactions on Multimedia Computing, Communications, and Applications, 19(2s), 1\u201331.","journal-title":"ACM Transactions on Multimedia Computing, Communications, and Applications"},{"key":"2198_CR13","doi-asserted-by":"crossref","unstructured":"Geng, Z., Liang, L., Ding, T., & Zharkov, I. (2022). Rstt: Real-time spatial temporal transformer for space-time video super-resolution. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 17441\u201317451","DOI":"10.1109\/CVPR52688.2022.01692"},{"key":"2198_CR14","doi-asserted-by":"crossref","unstructured":"Harshavardhan, S., Gupta, S., & Venkatesh, K. S. (2013). Flutter shutter based motion deblurring in complex scenes. In: 2013 Annual IEEE India Conference (INDICON), IEEE, pp 1\u20136","DOI":"10.1109\/INDCON.2013.6726105"},{"key":"2198_CR15","doi-asserted-by":"crossref","unstructured":"Hitomi, Y., Gu, J., Gupta, M., Mitsunaga, T., & Nayar, S. K. (2011). Video from a single coded exposure photograph using a learned over-complete dictionary. In: 2011 International Conference on Computer Vision, IEEE, pp 287\u2013294","DOI":"10.1109\/ICCV.2011.6126254"},{"key":"2198_CR16","doi-asserted-by":"crossref","unstructured":"Jeon, H. G., Lee, J. Y., Han, Y., Kim, S.J., & Kweon, I. S. (2015). Complementary sets of shutter sequences for motion deblurring. In: 2015 IEEE International Conference on Computer Vision (ICCV), IEEE, pp 3541\u20133549","DOI":"10.1109\/ICCV.2015.404"},{"issue":"2","key":"2198_CR17","doi-asserted-by":"publisher","first-page":"269","DOI":"10.1007\/s11263-016-0976-4","volume":"123","author":"HG Jeon","year":"2017","unstructured":"Jeon, H. G., Lee, J. Y., Han, Y., Kim, S. J., & Kweon, I. S. (2017). Generating fluttering patterns with low autocorrelation for coded exposure imaging. International Journal of Computer Vision, 123(2), 269\u2013286.","journal-title":"International Journal of Computer Vision"},{"key":"2198_CR18","doi-asserted-by":"crossref","unstructured":"Jin, M., Meishvili, G., & Favaro, P. (2018). Learning to extract a video sequence from a single motion-blurred image. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 6334\u20136342","DOI":"10.1109\/CVPR.2018.00663"},{"key":"2198_CR19","doi-asserted-by":"crossref","unstructured":"Jin, M., Hu. Z., & Favaro, P. (2019). Learning to extract flawless slow motion from blurry videos. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, pp 8104\u20138113","DOI":"10.1109\/CVPR.2019.00830"},{"key":"2198_CR20","first-page":"852","volume":"34","author":"T Karras","year":"2021","unstructured":"Karras, T., Aittala, M., Laine, S., H\u00e4rk\u00f6nen, E., Hellsten, J., Lehtinen, J., & Aila, T. (2021). Alias-free generative adversarial networks. Advances in Neural Information Processing Systems, 34, 852\u2013863.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2198_CR21","doi-asserted-by":"crossref","unstructured":"Ke, J., Wang, Q., Wang, Y., Milanfar, P., & Yang, F. (2021). MUSIQ: Multi-scale Image Quality Transformer. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), IEEE, pp 5128\u20135137","DOI":"10.1109\/ICCV48922.2021.00510"},{"issue":"12","key":"2198_CR22","doi-asserted-by":"publisher","first-page":"9396","DOI":"10.1109\/TPAMI.2021.3126387","volume":"44","author":"C Li","year":"2022","unstructured":"Li, C., Guo, C., Han, L., Jiang, J., Cheng, M. M., Gu, J., & Loy, C. C. (2022). Low-light image and video enhancement using deep learning: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(12), 9396\u20139416. https:\/\/doi.org\/10.1109\/TPAMI.2021.3126387","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"issue":"4","key":"2198_CR23","doi-asserted-by":"publisher","first-page":"700","DOI":"10.1109\/JSTSP.2022.3164524","volume":"16","author":"D Li","year":"2022","unstructured":"Li, D., Bian, L., & Zhang, J. (2022). High-speed large-scale imaging using frame decomposition from intrinsic multiplexing of motion. IEEE Journal of Selected Topics in Signal Processing, 16(4), 700\u2013712.","journal-title":"IEEE Journal of Selected Topics in Signal Processing"},{"key":"2198_CR24","doi-asserted-by":"crossref","unstructured":"Li, Z., Wang, M., Pi, H., Xu, K., Mei, J., & Liu, Y. (2022c). E-NeRV: Expedite neural video representation with disentangled spatial-temporal context. In: Computer Vision\u2014ECCV 2022, Springer Nature Switzerland, pp 267\u2013284","DOI":"10.1007\/978-3-031-19833-5_16"},{"key":"2198_CR25","doi-asserted-by":"crossref","unstructured":"Lin, S., Zhang, J., Pan, J., Jiang, Z., Zou, D., Wang, Y., Chen, J., & Ren. J. (2020). Learning event-driven video deblurring and interpolation. In: Computer Vision\u2014ECCV 2020, Springer International Publishing, pp 695\u2013710","DOI":"10.1007\/978-3-030-58598-3_41"},{"issue":"2","key":"2198_CR26","doi-asserted-by":"publisher","first-page":"248","DOI":"10.1109\/TPAMI.2013.129","volume":"36","author":"D Liu","year":"2014","unstructured":"Liu, D., Gu, J., Hitomi, Y., Gupta, M., Mitsunaga, T., & Nayar, S. K. (2014). Efficient space-time sampling with pixel-wise coded exposure for high-speed imaging. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36(2), 248\u2013260.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"issue":"9","key":"2198_CR27","doi-asserted-by":"publisher","first-page":"10526","DOI":"10.1364\/OE.21.010526","volume":"21","author":"P Llull","year":"2013","unstructured":"Llull, P., Liao, X., Yuan, X., Yang, J., Kittle, D., Carin, L., Sapiro, G., & Brady, D. J. (2013). Coded aperture compressive temporal imaging. Optics Express, 21(9), 10526\u201310545.","journal-title":"Optics Express"},{"key":"2198_CR28","unstructured":"Loshchilov, I., & Hutter, F. (2017). SGDR: Stochastic gradient descent with warm restarts. In: 2017 International Conference on Learning Representations (ICLR), p\u00a01"},{"key":"2198_CR29","doi-asserted-by":"crossref","unstructured":"Mai, L., & Liu, F. (2022). Motion-adjustable neural implicit video representation. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 10738\u201310747","DOI":"10.1109\/CVPR52688.2022.01047"},{"key":"2198_CR30","doi-asserted-by":"crossref","unstructured":"McCloskey, S. (2010). Velocity-dependent shutter sequences for motion deblurring. In: Computer Vision\u2014ECCV 2010, Springer, pp 309\u2013322","DOI":"10.1007\/978-3-642-15567-3_23"},{"issue":"10","key":"2198_CR31","doi-asserted-by":"publisher","first-page":"2071","DOI":"10.1109\/TPAMI.2012.108","volume":"34","author":"S McCloskey","year":"2012","unstructured":"McCloskey, S., Ding, Y., & Yu, J. (2012). Design and estimation of coded exposure point spread functions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34(10), 2071\u20132077.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"2198_CR32","doi-asserted-by":"crossref","unstructured":"Mildenhall, B., Srinivasan, P. P., Tancik, M., Barron, J. T., Ramamoorthi, R., & Ng, R. (2020). NeRF: Representing scenes as neural radiance fields for view synthesis. In: Computer vision\u2013ECCV 2020, Springer International Publishing, pp 405\u2013421","DOI":"10.1007\/978-3-030-58452-8_24"},{"key":"2198_CR33","doi-asserted-by":"crossref","unstructured":"Nah, S., Kim, T. H., Lee, K. M. (2017). Deep multi-scale convolutional neural network for dynamic scene deblurring. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, pp 257\u2013265","DOI":"10.1109\/CVPR.2017.35"},{"key":"2198_CR34","unstructured":"Nah, S., Son, S., Lee, J., & Lee, K. M. (2021). Clean images are hard to reblur: Exploiting the ill-posed inverse task for dynamic scene deblurring. In: 2021 International Conference on Learning Representations (ICLR)."},{"key":"2198_CR35","doi-asserted-by":"crossref","unstructured":"Pan, L., Scheerlinck, C., Yu, X., Hartley, R., Liu, M., & Dai, Y. (2019). Bringing a blurry frame alive at high frame-rate with an event camera. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 6820\u20136829","DOI":"10.1109\/CVPR.2019.00698"},{"issue":"1","key":"2198_CR36","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1007\/s00371-020-02016-y","volume":"38","author":"AS Parihar","year":"2022","unstructured":"Parihar, A. S., Varshney, D., Pandya, K., & Aggarwal, A. (2022). A comprehensive survey on video frame interpolation techniques. The Visual Computer, 38(1), 295\u2013319.","journal-title":"The Visual Computer"},{"key":"2198_CR37","first-page":"8024","volume":"32","author":"A Paszke","year":"2019","unstructured":"Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., & Desmaison, A. (2019). PyTorch: An imperative style, high-performance deep learning library. Advances in Neural Information Processing Systems, 32, 8024\u20138035.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2198_CR38","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1017\/S0962492900002919","volume":"8","author":"A Pinkus","year":"1999","unstructured":"Pinkus, A. (1999). Approximation theory of the MLP model in neural networks. Acta Numerica, 8, 143\u2013195.","journal-title":"Acta Numerica"},{"key":"2198_CR39","doi-asserted-by":"crossref","unstructured":"Purohit, K., Shah, A., & Rajagopalan, A. N. (2019). Bringing alive blurred moments. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 6830\u20136839","DOI":"10.1109\/CVPR.2019.00699"},{"key":"2198_CR40","doi-asserted-by":"crossref","unstructured":"Qiu, J., Wang, X., Maybank, S. J., & Tao, D. (2019). World From Blur. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, pp 8485\u20138496","DOI":"10.1109\/CVPR.2019.00869"},{"issue":"3","key":"2198_CR41","doi-asserted-by":"publisher","first-page":"795","DOI":"10.1145\/1141911.1141957","volume":"25","author":"R Raskar","year":"2006","unstructured":"Raskar, R., Agrawal, A., & Tumblin, J. (2006). Coded exposure photography: Motion deblurring using fluttered shutter. ACM Transactions on Graphics, 25(3), 795\u2013804.","journal-title":"ACM Transactions on Graphics"},{"issue":"6","key":"2198_CR42","doi-asserted-by":"publisher","first-page":"5317","DOI":"10.1007\/s10462-022-10302-5","volume":"56","author":"C Rota","year":"2023","unstructured":"Rota, C., Buzzelli, M., Bianco, S., & Schettini, R. (2023). Video restoration based on deep learning: A comprehensive survey. Artificial Intelligence Review, 56(6), 5317\u20135364.","journal-title":"Artificial Intelligence Review"},{"key":"2198_CR43","doi-asserted-by":"crossref","unstructured":"Rozumnyi, D., Oswald, M. R., Ferrari, V., Matas, J., & Pollefeys, M. (2021). DeFMO: Deblurring and shape recovery of fast moving objects. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 3456\u20133465","DOI":"10.1109\/CVPR46437.2021.00346"},{"key":"2198_CR44","doi-asserted-by":"publisher","first-page":"1051","DOI":"10.1109\/TCI.2022.3226947","volume":"8","author":"Y Sanghvi","year":"2022","unstructured":"Sanghvi, Y., Gnanasambandam, A., Mao, Z., & Chan, S. H. (2022). Photon-limited blind deconvolution using unsupervised iterative kernel estimation. IEEE Transactions on Computational Imaging, 8, 1051\u20131062.","journal-title":"IEEE Transactions on Computational Imaging"},{"key":"2198_CR45","doi-asserted-by":"crossref","unstructured":"Shangguan, W., Sun, Y., Gan, W., & Kamilov, U. S. (2022). Learning cross-video neural representations for high-quality frame interpolation. In: Computer Vision\u2013ECCV 2022, Springer Nature Switzerland, pp 511\u2013528.","DOI":"10.1007\/978-3-031-19784-0_30"},{"key":"2198_CR46","doi-asserted-by":"crossref","unstructured":"Shedligeri, P. S. A., & Mitra, K. (2021). A unified framework for compressive video recovery from coded exposure techniques. In: 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), IEEE, pp 1599\u20131608","DOI":"10.1109\/WACV48630.2021.00164"},{"key":"2198_CR47","doi-asserted-by":"crossref","unstructured":"Shen, W., Bao, W., Zhai, G., Chen, L., Min, X., & Gao, Z. (2020). Blurry video frame interpolation. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 5114\u20135123","DOI":"10.1109\/CVPR42600.2020.00516"},{"key":"2198_CR48","first-page":"7537","volume":"33","author":"M Tancik","year":"2020","unstructured":"Tancik, M., Srinivasan, P., Mildenhall, B., Fridovich-Keil, S., Raghavan, N., Singhal, U., Ramamoorthi, R., Barron, J., & Ng, R. (2020). Fourier features let networks learn high frequency functions in low dimensional domains. Advances in Neural Information Processing Systems, Curran Associates Inc, 33, 7537\u20137547.","journal-title":"Advances in Neural Information Processing Systems, Curran Associates Inc"},{"issue":"4","key":"2198_CR49","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A., Sheikh, H., & Simoncelli, E. (2004). Image quality assessment: From error visibility to structural similarity. IEEE Transactions on Image Processing, 13(4), 600\u2013612.","journal-title":"IEEE Transactions on Image Processing"},{"key":"2198_CR50","doi-asserted-by":"crossref","unstructured":"Xie, X., Zhou, P., Li, H., Lin, Z., & Yan, S. (2023). Adan: Adaptive nesterov momentum algorithm for faster optimizing deep models. arXiv preprint arXiv:2208.06677","DOI":"10.1109\/TPAMI.2024.3423382"},{"key":"2198_CR51","doi-asserted-by":"crossref","unstructured":"Yang, R., Xiao, T., Cheng, Y., Cao, Q., Qu, J., Suo, J., & Dai, Q. (2022). SCI: A spectrum concentrated implicit neural compression for biomedical data. arXiv preprint arXiv:2209.15180","DOI":"10.1609\/aaai.v37i4.25602"},{"key":"2198_CR52","doi-asserted-by":"crossref","unstructured":"Zhang, K., Luo, W., Stenger, B., Ren, W., Ma, L., & Li, H. (2020a). Every moment matters: Detail-aware networks to bring a blurry image alive. In: 28th ACM International Conference on Multimedia, ACM, pp 384\u2013392.","DOI":"10.1145\/3394171.3413929"},{"issue":"9","key":"2198_CR53","doi-asserted-by":"publisher","first-page":"2103","DOI":"10.1007\/s11263-022-01633-5","volume":"130","author":"K Zhang","year":"2022","unstructured":"Zhang, K., Ren, W., Luo, W., Lai, W. S., Stenger, B., Yang, M. H., & Li, H. (2022). Deep image deblurring: A survey. International Journal of Computer Vision, 130(9), 2103\u20132130.","journal-title":"International Journal of Computer Vision"},{"key":"2198_CR54","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A. A., Shechtman, E., & Wang, O. (2018). The unreasonable effectiveness of deep features as a perceptual metric. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, pp 586\u2013595.","DOI":"10.1109\/CVPR.2018.00068"},{"issue":"1","key":"2198_CR55","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1109\/TCSVT.2018.2886771","volume":"30","author":"W Zhang","year":"2020","unstructured":"Zhang, W., Ma, K., Yan, J., Deng, D., & Wang, Z. (2020). Blind Image Quality Assessment Using a Deep Bilinear Convolutional Neural Network. IEEE Transactions on Circuits and Systems for Video Technology, 30(1), 36\u201347.","journal-title":"IEEE Transactions on Circuits and Systems for Video Technology"},{"issue":"11","key":"2198_CR56","doi-asserted-by":"publisher","first-page":"2277","DOI":"10.1364\/PRJ.435256","volume":"9","author":"Z Zhang","year":"2021","unstructured":"Zhang, Z., Deng, C., Liu, Y., Yuan, X., Suo, J., & Dai, Q. (2021). Ten-mega-pixel snapshot compressive imaging with a hybrid coded aperture. Photonics Research, 9(11), 2277\u20132287.","journal-title":"Photonics Research"},{"key":"2198_CR57","doi-asserted-by":"publisher","first-page":"1390","DOI":"10.1109\/TIP.2023.3244417","volume":"32","author":"Z Zhang","year":"2023","unstructured":"Zhang, Z., Cheng, Y., Suo, J., Bian, L., & Dai, Q. (2023). INFWIDE: Image and feature space wiener deconvolution network for non-blind image deblurring in low-light conditions. IEEE Transactions on Image Processing, 32, 1390\u20131402.","journal-title":"IEEE Transactions on Image Processing"},{"issue":"10","key":"2198_CR58","doi-asserted-by":"publisher","first-page":"1678","DOI":"10.1364\/PRJ.489989","volume":"11","author":"Z Zhang","year":"2023","unstructured":"Zhang, Z., Dong, K., Suo, J., & Dai, Q. (2023). Deep coded exposure: End-to-end co-optimization of flutter shutter and deblurring processing for general motion blur removal. Photonics Research, 11(10), 1678.","journal-title":"Photonics Research"},{"key":"2198_CR59","doi-asserted-by":"crossref","unstructured":"Zhong, Z., Sun, X., Wu, Z., Zheng, Y., Lin, S., & Sato, I. (2022). Animation from\u00a0Blur: Multi-modal blur decomposition with motion guidance. In: Computer Vision\u2013ECCV 2022, Springer Nature Switzerland, pp 599\u2013615","DOI":"10.1007\/978-3-031-19800-7_35"},{"key":"2198_CR60","doi-asserted-by":"crossref","unstructured":"Zuckerman, L. P., Naor, E., Pisha, G., Bagon, S., & Irani, M. (2020). Across scales and across dimensions: Temporal super-resolution using deep internal learning. In: Computer Vision\u2013ECCV 2020, Springer International Publishing, pp 52\u201368.","DOI":"10.1007\/978-3-030-58571-6_4"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-024-02198-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-024-02198-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-024-02198-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,24]],"date-time":"2025-02-24T10:04:12Z","timestamp":1740391452000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-024-02198-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,30]]},"references-count":60,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,3]]}},"alternative-id":["2198"],"URL":"https:\/\/doi.org\/10.1007\/s11263-024-02198-1","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,8,30]]},"assertion":[{"value":"31 May 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 July 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 August 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declaration"}},{"value":"The authors have no relevant financial or non-financial interests to disclose.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}