{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,9]],"date-time":"2025-05-09T14:48:12Z","timestamp":1746802092210,"version":"3.37.3"},"reference-count":59,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2022,8,19]],"date-time":"2022-08-19T00:00:00Z","timestamp":1660867200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,8,19]],"date-time":"2022-08-19T00:00:00Z","timestamp":1660867200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100003213","name":"Beijing Municipal Education Commission","doi-asserted-by":"publisher","award":["KJZD20191000402"],"award-info":[{"award-number":["KJZD20191000402"]}],"id":[{"id":"10.13039\/501100003213","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["51827813","61472029"],"award-info":[{"award-number":["51827813","61472029"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2023,5]]},"DOI":"10.1007\/s10489-022-03787-8","type":"journal-article","created":{"date-parts":[[2022,8,19]],"date-time":"2022-08-19T12:02:46Z","timestamp":1660910566000},"page":"10535-10547","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Enhanced spatial-temporal freedom for video frame interpolation"],"prefix":"10.1007","volume":"53","author":[{"given":"Hao-Dong","family":"Li","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4226-4368","authenticated-orcid":false,"given":"Hui","family":"Yin","sequence":"additional","affiliation":[]},{"given":"Zhi-Hao","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Hua","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,8,19]]},"reference":[{"key":"3787_CR1","doi-asserted-by":"crossref","unstructured":"Mahajan D, Huang FC, Matusik W, Ramamoorthi R, Belhumeur P (2009) Moving gradients: a path-based method for plausible image interpolation. ACM Transactions on Graphics (TOG), pp 1\u201311","DOI":"10.1145\/1531326.1531348"},{"key":"3787_CR2","doi-asserted-by":"crossref","unstructured":"Liu Z, Yeh RA, Tang X, Liu Y, Agarwala A (2017) Video frame synthesis using deep voxel flow. In: 2017 IEEE International conference on computer vision (ICCV), pp 4473\u20134481","DOI":"10.1109\/ICCV.2017.478"},{"key":"3787_CR3","doi-asserted-by":"crossref","unstructured":"Liu Y, Liao YT, Lin YY, Chuang YY (2019) Deep video frame interpolation using cyclic frame generation. In: AAAI","DOI":"10.1609\/aaai.v33i01.33018794"},{"key":"3787_CR4","doi-asserted-by":"publisher","unstructured":"Bao W, Lai WS, Zhang X, Gao Z, Yang MH (2019) Memc-net: motion estimation and motion compensation driven neural network for video interpolation and enhancement. IEEE Trans Pattern Anal Mach Intell. https:\/\/doi.org\/10.1109\/TPAMI.2019.2941941","DOI":"10.1109\/TPAMI.2019.2941941"},{"key":"3787_CR5","unstructured":"Myungsub C, Choi J, Baik S, Kim T, Lee KM (2020) Scene adaptive video frame interpolation via meta-learning. In: 2020 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 9441\u20139450"},{"key":"3787_CR6","doi-asserted-by":"crossref","unstructured":"Tulyakov S, Gehrig D, Georgoulis S, Erbach J, Gehrig M, Li Y, Scaramuzza D (2021) Time lens: Event-Based video frame interpolation. In: 2021 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 16155\u201316164","DOI":"10.1109\/CVPR46437.2021.01589"},{"key":"3787_CR7","doi-asserted-by":"crossref","unstructured":"Bao W, Zhang X, Chen L, Ding L, Gao Z (2018) High-order model and dynamic filtering for frame rate up-conversion. IEEE Trans Image Process, pp 3813\u20133826","DOI":"10.1109\/TIP.2018.2825100"},{"key":"3787_CR8","doi-asserted-by":"crossref","unstructured":"Jiang H, Sun D, Jampani V, Yang MH, Learned-Miller E, Kautz J (2018) Super slomo: high quality estimation of multiple intermediate frames for video interpolation. In: 2018 IEEE\/CVF Conference on computer vision and pattern recognition, pp 9000\u20139008","DOI":"10.1109\/CVPR.2018.00938"},{"key":"3787_CR9","doi-asserted-by":"crossref","unstructured":"Flynn J, Neulander I, Philbin J, Snavely N (2016) Deepstereo: Learning to predict new views from the world\u2019s imagery. In: 2016 IEEE Conference on computer vision and pattern recognition (CVPR), pp 5515\u20135524","DOI":"10.1109\/CVPR.2016.595"},{"key":"3787_CR10","doi-asserted-by":"crossref","unstructured":"Zhou T, Tulsiani S, Sun W, Malik J, Efros AA (2016) View synthesis by appearance flow. In: European conference on computer vision (ECCV), pp 286\u2013301","DOI":"10.1007\/978-3-319-46493-0_18"},{"key":"3787_CR11","unstructured":"Kalluri T, Pathak D, Chandraker M, Tran D (2020) Flavr: Flow-agnostic video representations for fast frame interpolation. arXiv:2012.08512"},{"key":"3787_CR12","doi-asserted-by":"crossref","unstructured":"Wu J, Yuen C, Cheung NM, Chen J, Chen CW (2015) Modeling and optimization of high frame rate video transmission over wireless networks. IEEE Trans Wirel Commun, pp 2713\u2013 2726","DOI":"10.1109\/TWC.2015.2509063"},{"key":"3787_CR13","doi-asserted-by":"crossref","unstructured":"Parihar AS, Varshney D, Pandya K, Aggarwal A (2021) A comprehensive survey on video frame interpolation techniques. Vis Comput, pp 1\u201325","DOI":"10.1007\/s00371-020-02016-y"},{"key":"3787_CR14","doi-asserted-by":"crossref","unstructured":"Niklaus S, Mai L, Liu F (2017) Video frame interpolation via adaptive convolution. In: 2017 IEEE Conference on computer vision and pattern recognition (CVPR), pp 2270\u20132279","DOI":"10.1109\/CVPR.2017.244"},{"key":"3787_CR15","doi-asserted-by":"crossref","unstructured":"Niklaus S, Mai L, Liu F (2017) Video frame interpolation via adaptive separable convolution. In: 2017 IEEE International conference on computer vision (ICCV), pp 261\u2013270","DOI":"10.1109\/ICCV.2017.37"},{"key":"3787_CR16","doi-asserted-by":"crossref","unstructured":"Lee H, Kim T, Chung TY, Pak D, Ban Y, Lee S (2020) Adacof: adaptive collaboration of flows for video frame interpolation. In: 2020 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 5315\u20135324","DOI":"10.1109\/CVPR42600.2020.00536"},{"key":"3787_CR17","doi-asserted-by":"crossref","unstructured":"Dai J, Qi H, Xiong Y, Li Y, Zhang G, Hu H, Wei Y (2017) Deformable convolutional networks. In: 2017 IEEE International conference on computer vision (ICCV), pp 764\u2013773","DOI":"10.1109\/ICCV.2017.89"},{"key":"3787_CR18","doi-asserted-by":"crossref","unstructured":"Zhu X, Hu H, Lin S, Dai J (2019) Deformable convnets v2: More deformable, better results. In: 2019 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 9308\u20139316","DOI":"10.1109\/CVPR.2019.00953"},{"key":"3787_CR19","doi-asserted-by":"crossref","unstructured":"Shi Z, Liu X, Shi K, Dai L, Chen J (2021) Video frame interpolation via generalized deformable convolution. IEEE transactions on multimedia","DOI":"10.1109\/TMM.2021.3052419"},{"key":"3787_CR20","doi-asserted-by":"crossref","unstructured":"Chi Z, Mohammadi Nasiri R, Liu Z, Lu J, Tang J, Plataniotis KN (2020) All at once: Temporally adaptive multi-frame interpolation with advanced motion modeling. In: 2020 European conference on computer vision (ECCV), pp 107\u2013123","DOI":"10.1007\/978-3-030-58583-9_7"},{"key":"3787_CR21","doi-asserted-by":"crossref","unstructured":"Liu Y, Xie L, Siyao L, Sun W, Qiao Y, Dong C (2020) Enhanced quadratic video interpolation. In: European conference on computer vision (ECCV), pp 41\u201356","DOI":"10.1007\/978-3-030-66823-5_3"},{"key":"3787_CR22","unstructured":"Xu X, Siyao L, Sun W, Yin Q, Yang MH (2019) Quadratic video interpolation. arXiv:1911.00627"},{"key":"3787_CR23","doi-asserted-by":"crossref","unstructured":"Wang X, Jin Y, Li C, Cen Y, Li Y (2022) VSLN: View-aware Sphere learning network for cross-view vehicle re-identification. Int J Intell Syst, pp 1\u201321","DOI":"10.1002\/int.22857"},{"key":"3787_CR24","doi-asserted-by":"crossref","unstructured":"Park J, Ko K, Lee C, Kim CS (2020) Bmbc: Bilateral motion estimation with bilateral cost volume for video interpolation. In: 2020 European conference on computer vision (ECCV), pp 109\u2013125","DOI":"10.1007\/978-3-030-58568-6_7"},{"key":"3787_CR25","doi-asserted-by":"crossref","unstructured":"Siyao L, Zhao S, Yu W, Sun W, Metaxas D, Loy CC, Liu Z (2021) Deep animation video interpolation in the wild. In: 2021 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 6587\u20136595","DOI":"10.1109\/CVPR46437.2021.00652"},{"key":"3787_CR26","doi-asserted-by":"crossref","unstructured":"Teed Z, Deng J (2020) Raft: Recurrent all-pairs field transforms for optical flow. In: 2020 European conference on computer vision (ECCV), pp 402\u2013419","DOI":"10.1007\/978-3-030-58536-5_24"},{"key":"3787_CR27","doi-asserted-by":"crossref","unstructured":"Zhang H, Zhao Y, Wang R (2020) A flexible recurrent residual pyramid network for video frame interpolation. In: 2020 European conference on computer vision (ECCV), pp 474\u2013491","DOI":"10.1007\/978-3-030-58595-2_29"},{"key":"3787_CR28","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention, pp 234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"3787_CR29","doi-asserted-by":"crossref","unstructured":"Bao W, Lai WS, Ma C, Zhang X, Gao Z, Yang MH (2019) Depth aware video frame interpolation. In: 2019 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 3698\u20133707","DOI":"10.1109\/CVPR.2019.00382"},{"key":"3787_CR30","doi-asserted-by":"crossref","unstructured":"Niklaus S, Liu F (2020) Softmax splatting for video frame interpolation. In: 2020 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 5436\u20135445","DOI":"10.1109\/CVPR42600.2020.00548"},{"key":"3787_CR31","doi-asserted-by":"crossref","unstructured":"Sun D, Yang X, Liu MY, Kautz J (2018) Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume. In: 2018 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 8934\u20138943","DOI":"10.1109\/CVPR.2018.00931"},{"key":"3787_CR32","doi-asserted-by":"crossref","unstructured":"Sim H, Oh J, Kim M (2021) XVFI: Extreme video frame interpolation. In: 2021 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 14489\u201314498","DOI":"10.1109\/ICCV48922.2021.01422"},{"key":"3787_CR33","doi-asserted-by":"crossref","unstructured":"Lee S, Choi N, Choi WI (2022) Enhanced correlation matching based video frame interpolation. In: 2022 Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 2839\u20132847","DOI":"10.1109\/WACV51458.2022.00318"},{"key":"3787_CR34","doi-asserted-by":"crossref","unstructured":"Ding T, Liang L, Zhu Z, Zharkov I (2021) CDFI: Compression-Driven Network design for frame interpolation. In: 2021 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 8001\u20138011","DOI":"10.1109\/CVPR46437.2021.00791"},{"key":"3787_CR35","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1007\/s10489-021-02246-0","volume":"52","author":"Y Zhang","year":"2022","unstructured":"Zhang Y, Sun Y, Liu S (2022) Deformable and residual convolutional network for image super-resolution. Appl Intell 52:295\u2013304","journal-title":"Appl Intell"},{"issue":"4","key":"3787_CR36","doi-asserted-by":"publisher","first-page":"1100","DOI":"10.1007\/s10489-019-01603-4","volume":"50","author":"M Lu","year":"2020","unstructured":"Lu M, Hu Y, Lu X (2020) Driver action recognition using deformable and dilated faster r-CNN with optimized region proposals. Appl Intell 50(4):1100\u20131111","journal-title":"Appl Intell"},{"issue":"1","key":"3787_CR37","doi-asserted-by":"publisher","first-page":"427","DOI":"10.1007\/s10489-020-01842-w","volume":"51","author":"YB Liu","year":"2021","unstructured":"Liu YB, Jia RS, Liu QM, Zhang XL, Sun HM (2021) Crowd counting method based on the self-attention residual network. Appl Intell 51(1):427\u2013440","journal-title":"Appl Intell"},{"key":"3787_CR38","doi-asserted-by":"crossref","unstructured":"Cheng X, Chen Z (2020) Video frame interpolation via deformable separable convolution. In: AAAI","DOI":"10.1609\/aaai.v34i07.6634"},{"key":"3787_CR39","doi-asserted-by":"crossref","unstructured":"Gui S, Wang C, Chen Q, Tao D (2020) Featureflow: Robust video interpolation via structure-to-texture generation. In: 2020 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 14004\u201314013","DOI":"10.1109\/CVPR42600.2020.01402"},{"key":"3787_CR40","doi-asserted-by":"crossref","unstructured":"Yuan M, Dai Q (2021) A novel deep pixel restoration video prediction algorithm integrating attention mechanism. Appl Intell, pp 1\u201319","DOI":"10.1007\/s10489-021-02631-9"},{"key":"3787_CR41","doi-asserted-by":"crossref","unstructured":"Jing B, Ding H, Yang Z, Li B, Bao L (2021) Video prediction: a step-by-step improvement of a video synthesis network. Appl Intell, pp 1\u201313","DOI":"10.1007\/s10489-021-02500-5"},{"key":"3787_CR42","doi-asserted-by":"crossref","unstructured":"Wang X, Jin Y, Cen Y, Lang C, Li Y (2021) PST-NET: Point cloud sampling via Point-Based transformer. In: International conference on image and graphics, pp 57\u201369","DOI":"10.1007\/978-3-030-87361-5_5"},{"key":"3787_CR43","doi-asserted-by":"crossref","unstructured":"Kumar N, Sukavanam N (2020) An improved CNN framework for detecting and tracking human body in unconstraint environment. Knowledge-Based Systems, pp 193, 105198","DOI":"10.1016\/j.knosys.2019.105198"},{"key":"3787_CR44","doi-asserted-by":"crossref","unstructured":"Niklaus S, Liu F (2018) Context-aware synthesis for video frame interpolation. In: 2018 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 1701\u20131710","DOI":"10.1109\/CVPR.2018.00183"},{"key":"3787_CR45","doi-asserted-by":"crossref","unstructured":"Hou Q, Zhou D, Feng J (2021) Coordinate attention for efficient mobile network design. In: 2021 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 13713\u201313722","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"3787_CR46","doi-asserted-by":"crossref","unstructured":"Odena A, Dumoulin V, Olah C (2016) Deconvolution and checkerboard artifacts. Distill e3","DOI":"10.23915\/distill.00003"},{"key":"3787_CR47","doi-asserted-by":"crossref","unstructured":"Fourure D, Emonet R, Fromont E, Muselet D, Tremeau A, Wolf C (2017) Residual conv-deconv grid network for semantic segmentation. arXiv:1707.07958","DOI":"10.5244\/C.31.181"},{"key":"3787_CR48","doi-asserted-by":"crossref","unstructured":"Zhang Y, Tian Y, Kong Y, Zhong B, Fu Y (2018) Residual dense network for image super-resolution. In: 2018 IEEE Conference on computer vision and pattern recognition (CVPR), pp 2472\u20132481","DOI":"10.1109\/CVPR.2018.00262"},{"key":"3787_CR49","unstructured":"Wang X, Jin Y, Cen Y, Wang T, Tang B, Li Y (2022) LighTN: Light-weight Transformer Network for Performance-overhead Tradeoff in Point Cloud Downsampling. arXiv:2202.06263"},{"key":"3787_CR50","unstructured":"Kingma DP, Ba J (2014) Adam: A method for stochastic optimization. arXiv:1412.6980"},{"key":"3787_CR51","doi-asserted-by":"publisher","first-page":"1106","DOI":"10.1007\/s11263-018-01144-2","volume":"127","author":"T Xue","year":"2018","unstructured":"Xue T, Chen B, Wu J, Wei D, Freeman W (2018) Video enhancement with task-oriented flow. Int J Comput Vis 127:1106\u20131125","journal-title":"Int J Comput Vis"},{"key":"3787_CR52","doi-asserted-by":"crossref","unstructured":"Nah S, Hyun Kim T, Mu Lee K (2017) Deep multi-scale convolutional neural network for dynamic scene deblurring. In: 2017 IEEE Conference on computer vision and pattern recognition, pp 3883\u20133891","DOI":"10.1109\/CVPR.2017.35"},{"key":"3787_CR53","doi-asserted-by":"crossref","unstructured":"Su S, Delbracio M, Wang J, Sapiro G, Heidrich W, Wang O (2017) Deep video deblurring for hand-held cameras. In: 2017 IEEE Conference on computer vision and pattern recognition (CVPR), pp 237\u2013246","DOI":"10.1109\/CVPR.2017.33"},{"key":"3787_CR54","unstructured":"Soomro K, Zamir A, Shah M (2012) Ucf101: A dataset of 101 human actions classes from videos in the wild. arXiv:1212.0402"},{"key":"3787_CR55","doi-asserted-by":"crossref","unstructured":"Wang Z, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing, pp 600\u2013612","DOI":"10.1109\/TIP.2003.819861"},{"key":"3787_CR56","unstructured":"Nilsson J, Akenine-m\u00f6ller T (2020) Understanding ssim. arXiv:2006.13846"},{"key":"3787_CR57","doi-asserted-by":"crossref","unstructured":"Zhang D, Lei W, Zhang W, Chen X (2021) Flow-based frame interpolation networks combined with occlusion-aware mask estimation. IET Image Processing, pp 4579\u20134587","DOI":"10.1049\/iet-ipr.2020.0586"},{"key":"3787_CR58","doi-asserted-by":"crossref","unstructured":"Xiang X, Tian Y, Zhang Y, Fu Y, Allebach JP, Xu C (2021) Zooming SlowMo: An Efficient One-Stage Framework for Space-Time Video Super-Resolution. arXiv:2104.07473","DOI":"10.1109\/CVPR42600.2020.00343"},{"key":"3787_CR59","doi-asserted-by":"crossref","unstructured":"Xu G, Xu J, Li Z, Wang L, Sun X, Cheng MM (2021) Temporal modulation network for controllable Space-Time video Super-Resolution. In: 2021 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 6388\u20136397","DOI":"10.1109\/CVPR46437.2021.00632"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03787-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-022-03787-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03787-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,5,19]],"date-time":"2023-05-19T11:47:05Z","timestamp":1684496825000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-022-03787-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,8,19]]},"references-count":59,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2023,5]]}},"alternative-id":["3787"],"URL":"https:\/\/doi.org\/10.1007\/s10489-022-03787-8","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"type":"print","value":"0924-669X"},{"type":"electronic","value":"1573-7497"}],"subject":[],"published":{"date-parts":[[2022,8,19]]},"assertion":[{"value":"17 May 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 August 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}