{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,24]],"date-time":"2026-01-24T20:39:11Z","timestamp":1769287151331,"version":"3.49.0"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2025,4,23]],"date-time":"2025-04-23T00:00:00Z","timestamp":1745366400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,4,23]],"date-time":"2025-04-23T00:00:00Z","timestamp":1745366400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"\u00d6zye\u011fin University"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Machine Vision and Applications"],"published-print":{"date-parts":[[2025,5]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>High-spatio-temporal resolution (HSTR) video recording plays a crucial role in enhancing various imagery tasks that require fine-detailed information. State-of-the-art cameras provide this required high frame-rate and high spatial resolution together, albeit at a high cost. To alleviate this issue, this paper proposes a dual camera system for the generation of HSTR video using reference-based super-resolution (RefSR). One camera captures high spatial resolution low frame rate (HSLF) video while the other captures low spatial resolution high frame rate (LSHF) video simultaneously for the same scene. A novel deep learning architecture is proposed to fuse HSLF and LSHF video feeds and synthesize HSTR video frames. The proposed model combines optical flow estimation and (channel-wise and spatial) attention mechanisms to capture the fine motion and complex dependencies between frames of the two video feeds. Simulations show that the proposed model provides significant improvement over existing reference-based SR techniques in terms of PSNR and SSIM metrics. The method also exhibits sufficient frames per second (FPS) for aerial monitoring when deployed on a power-constrained drone equipped with dual cameras. The source code is publicly available at <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" xlink:href=\"https:\/\/github.com\/umutsuluhan\/HSTRNet\" ext-link-type=\"uri\">https:\/\/github.com\/umutsuluhan\/HSTRNet<\/jats:ext-link>.<\/jats:p>","DOI":"10.1007\/s00138-025-01685-w","type":"journal-article","created":{"date-parts":[[2025,4,23]],"date-time":"2025-04-23T06:18:23Z","timestamp":1745389103000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Hstr-net: reference based video super-resolution with dual cameras"],"prefix":"10.1007","volume":"36","author":[{"given":"H. Umut","family":"Suluhan","sequence":"first","affiliation":[]},{"given":"Abdullah Enes","family":"Doruk","sequence":"additional","affiliation":[]},{"given":"Hasan F.","family":"Ates","sequence":"additional","affiliation":[]},{"given":"Bahadir K.","family":"Gunturk","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,23]]},"reference":[{"issue":"10","key":"1685_CR1","doi-asserted-by":"publisher","first-page":"3275","DOI":"10.1109\/TPAMI.2020.2983371","volume":"43","author":"M Cheng","year":"2020","unstructured":"Cheng, M., Ma, Z., Asif, M.S., Xu, Y., Liu, H., Bao, W., Sun, J.: A dual camera system for high spatiotemporal resolution video acquisition. IEEE Trans. Pattern Anal. Mach. Intell. 43(10), 3275\u20133291 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"4","key":"1685_CR2","first-page":"1","volume":"20","author":"M Zhou","year":"2023","unstructured":"Zhou, M., Zhao, X., Luo, F., Luo, J., Pu, H., Xiang, T.: Robust rgb-t tracking via adaptive modality weight correlation filters and cross-modality learning. ACM Trans. Multimed. Comput. Commun. Appl. 20(4), 1\u201320 (2023)","journal-title":"ACM Trans. Multimed. Comput. Commun. Appl."},{"key":"1685_CR3","doi-asserted-by":"publisher","first-page":"650","DOI":"10.1109\/TMM.2023.3268867","volume":"26","author":"M Zhou","year":"2023","unstructured":"Zhou, M., Wu, X., Wei, X., Xiang, T., Fang, B., Kwong, S.: Low-light enhancement method based on a retinex model for structure preservation. IEEE Trans. Multimedia 26, 650\u2013662 (2023)","journal-title":"IEEE Trans. Multimedia"},{"issue":"2","key":"1685_CR4","doi-asserted-by":"publisher","first-page":"369","DOI":"10.1109\/TBC.2022.3215249","volume":"69","author":"M Zhou","year":"2022","unstructured":"Zhou, M., Lan, X., Wei, X., Liao, X., Mao, Q., Li, Y., Wu, C., Xiang, T., Fang, B.: An end-to-end blind image quality assessment method using a recurrent network and self-attention. IEEE Trans. Broadcast. 69(2), 369\u2013377 (2022)","journal-title":"IEEE Trans. Broadcast."},{"issue":"2","key":"1685_CR5","doi-asserted-by":"publisher","first-page":"396","DOI":"10.1109\/TBC.2022.3231101","volume":"69","author":"M Zhou","year":"2023","unstructured":"Zhou, M., Chen, L., Wei, X., Liao, X., Mao, Q., Wang, H., Pu, H., Luo, J., Xiang, T., Fang, B.: Perception-oriented u-shaped transformer network for 360-degree no-reference image quality assessment. IEEE Trans. Broadcast. 69(2), 396\u2013405 (2023)","journal-title":"IEEE Trans. Broadcast."},{"key":"1685_CR6","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-024-02338-7","author":"M Zhou","year":"2025","unstructured":"Zhou, M., Shen, W., Wei, X., Luo, J., Jia, F., Zhuang, X., Jia, W.: Blind image quality assessment Exploring content fidelity perceptibility via quality adversarial learning. Int. J. Comput. Vision (2025). https:\/\/doi.org\/10.1007\/s11263-024-02338-7","journal-title":"Int. J. Comput. Vision"},{"issue":"8","key":"1685_CR7","doi-asserted-by":"publisher","first-page":"1106","DOI":"10.1007\/s11263-018-01144-2","volume":"127","author":"T Xue","year":"2019","unstructured":"Xue, T., Chen, B., Wu, J., Wei, D., Freeman, W.T.: Video enhancement with task-oriented flow. Int. J. Comput. Vision (IJCV) 127(8), 1106\u20131125 (2019)","journal-title":"Int. J. Comput. Vision (IJCV)"},{"issue":"2","key":"1685_CR8","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1109\/TPAMI.2015.2439281","volume":"38","author":"C Dong","year":"2015","unstructured":"Dong, C., Loy, C.C., He, K., Tang, X.: Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 38(2), 295\u2013307 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1685_CR9","doi-asserted-by":"crossref","unstructured":"Kim, J., Lee, J.K., Lee, K.M.: Accurate image super-resolution using very deep convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1646\u20131654 (2016)","DOI":"10.1109\/CVPR.2016.182"},{"key":"1685_CR10","doi-asserted-by":"crossref","unstructured":"Lim, B., Son, S., Kim, H., Nah, S., Mu\u00a0Lee, K.: Enhanced deep residual networks for single image super-resolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 136\u2013144 (2017)","DOI":"10.1109\/CVPRW.2017.151"},{"key":"1685_CR11","doi-asserted-by":"crossref","unstructured":"Ledig, C., Theis, L., Husz\u00e1r, F., Caballero, J., Cunningham, A., Acosta, A., Aitken, A., Tejani, A., Totz, J., Wang, Z., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4681\u20134690 (2017)","DOI":"10.1109\/CVPR.2017.19"},{"key":"1685_CR12","unstructured":"Denton, E.L., Chintala, S., Fergus, R., et al.: Deep generative image models using a laplacian pyramid of adversarial networks. Advances in neural information processing systems 28 (2015)"},{"key":"1685_CR13","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, K., Wu, S., Gu, J., Liu, Y., Dong, C., Qiao, Y., Change\u00a0Loy, C.: Esrgan: Enhanced super-resolution generative adversarial networks. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops, (2018)","DOI":"10.1007\/978-3-030-11021-5_5"},{"issue":"11","key":"1685_CR14","doi-asserted-by":"publisher","first-page":"4544","DOI":"10.1109\/TIP.2012.2208977","volume":"21","author":"K Zhang","year":"2012","unstructured":"Zhang, K., Gao, X., Tao, D., Li, X.: Single image super-resolution with non-local means and steering kernel regression. IEEE Trans. Image Process. 21(11), 4544\u20134556 (2012). https:\/\/doi.org\/10.1109\/TIP.2012.2208977","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"1685_CR15","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1109\/38.988747","volume":"22","author":"WT Freeman","year":"2002","unstructured":"Freeman, W.T., Jones, T.R., Pasztor, E.C.: Example-based super-resolution. IEEE Comput. Graphics Appl. 22(2), 56\u201365 (2002)","journal-title":"IEEE Comput. Graphics Appl."},{"key":"1685_CR16","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Wang, Z., Lin, Z., Qi, H.: Image super-resolution by neural texture transfer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7982\u20137991 (2019)","DOI":"10.1109\/CVPR.2019.00817"},{"key":"1685_CR17","doi-asserted-by":"crossref","unstructured":"Shim, G., Park, J., Kweon, I.S.: Robust reference-based super-resolution with similarity-aware deformable convolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8425\u20138434 (2020)","DOI":"10.1109\/CVPR42600.2020.00845"},{"key":"1685_CR18","doi-asserted-by":"crossref","unstructured":"Yang, F., Yang, H., Fu, J., Lu, H., Guo, B.: Learning texture transformer network for image super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5791\u20135800 (2020)","DOI":"10.1109\/CVPR42600.2020.00583"},{"key":"1685_CR19","doi-asserted-by":"crossref","unstructured":"Lu, L., Li, W., Tao, X., Lu, J., Jia, J.: Masa-sr: Matching acceleration and spatial adaptation for reference-based image super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6368\u20136377 (2021)","DOI":"10.1109\/CVPR46437.2021.00630"},{"key":"1685_CR20","doi-asserted-by":"crossref","unstructured":"Lin, R., Xiao, N.: Residual channel attention connection network for reference-based image super-resolution. In: 2021 8th International Conference on Information, Cybernetics, and Computational Social Systems (ICCSS). IEEE, pp. 307\u2013313 (2021)","DOI":"10.1109\/ICCSS53909.2021.9722011"},{"key":"1685_CR21","doi-asserted-by":"crossref","unstructured":"Huang, Y., Zhang, X., Fu, Y., Chen, S., Zhang, Y., Wang, Y.-F., He, D.: Task decoupled framework for reference-based super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5931\u20135940 (2022)","DOI":"10.1109\/CVPR52688.2022.00584"},{"key":"1685_CR22","doi-asserted-by":"publisher","first-page":"54649","DOI":"10.1109\/ACCESS.2022.3174194","volume":"10","author":"X Liu","year":"2022","unstructured":"Liu, X., Li, J., Duan, T., Li, J., Wang, Y.: Dsma: Reference-based image super-resolution method based on dual-view supervised learning and multi-attention mechanism. IEEE Access 10, 54649\u201354659 (2022)","journal-title":"IEEE Access"},{"key":"1685_CR23","doi-asserted-by":"crossref","unstructured":"Zheng, H., Ji, M., Wang, H., Liu, Y., Fang, L.: Crossnet: An end-to-end reference-based super resolution network using cross-scale warping. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 88\u2013104 (2018)","DOI":"10.1007\/978-3-030-01231-1_6"},{"key":"1685_CR24","doi-asserted-by":"crossref","unstructured":"Jiang, Y., Chan, K.C., Wang, X., Loy, C.C., Liu, Z.: Robust reference-based super-resolution via c2-matching. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2103\u20132112 (2021)","DOI":"10.1109\/CVPR46437.2021.00214"},{"key":"1685_CR25","doi-asserted-by":"crossref","unstructured":"Wang, T., Xie, J., Sun, W., Yan, Q., Chen, Q.: Dual-camera super-resolution with aligned attention modules. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2001\u20132010 (2021)","DOI":"10.1109\/ICCV48922.2021.00201"},{"key":"1685_CR26","doi-asserted-by":"crossref","unstructured":"Xia, B., Tian, Y., Hang, Y., Yang, W., Liao, Q., Zhou, J.: Coarse-to-fine embedded patchmatch and multi-scale dynamic aggregation for reference-based super-resolution. arXiv preprint arXiv:2201.04358 (2022)","DOI":"10.1609\/aaai.v36i3.20180"},{"key":"1685_CR27","doi-asserted-by":"crossref","unstructured":"Lee, J., Lee, M., Cho, S., Lee, S.: Reference-based video super-resolution using multi-camera video triplets. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 17824\u201317833 (2022)","DOI":"10.1109\/CVPR52688.2022.01730"},{"key":"1685_CR28","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"1685_CR29","doi-asserted-by":"crossref","unstructured":"Suluhan, H.U., Ates, H.F., Gunturk, B.K.: Dual camera based high spatio-temporal resolution video generation for wide area surveillance. In: 2022 18th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS). IEEE, pp. 1\u20138 (2022)","DOI":"10.1109\/AVSS56176.2022.9959711"},{"key":"1685_CR30","unstructured":"Zhu, P., Wen, L., Du, D., Bian, X., Hu, Q., Ling, H.: Vision meets drones: past, present and future. 1(2), 8, arXiv preprint arXiv:2001.06303 (2020)"},{"key":"1685_CR31","unstructured":"Minor Area Motion Imagery (2013). https:\/\/www.sdms.afrl.af.mil\/index.php?collection=mami2013"},{"key":"1685_CR32","unstructured":"Jetson AGX Xavier Evaluation Board. Accessed: 2023-07-03 (2023). www.nvidia.com\/en-us\/autonomous-machines\/embedded-systems\/jetson-agx-xavier\/"},{"key":"1685_CR33","doi-asserted-by":"publisher","DOI":"10.1145\/3604937","author":"K Guo","year":"2023","unstructured":"Guo, K., Chen, L., Zhu, X., Kui, X., Zhang, J., Shi, H.: Double-layer search and adaptive pooling fusion for reference-based image super-resolution. ACM Trans. Multimedia Comput. Commun. Appl. (2023). https:\/\/doi.org\/10.1145\/3604937","journal-title":"ACM Trans. Multimedia Comput. Commun. Appl."},{"key":"1685_CR34","unstructured":"Mei, X., Yang, Y., Li, M., Huang, C., Zhang, K., Li\u00f3, P.: A feature reuse framework with texture-adaptive aggregation for reference-based super-resolution. arXiv preprint arXiv:2306.01500 (2023)"},{"key":"1685_CR35","doi-asserted-by":"crossref","unstructured":"Shim, G., Park, J., Kweon, I.S.: Robust reference-based super-resolution with similarity-aware deformable convolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8425\u20138434 (2020)","DOI":"10.1109\/CVPR42600.2020.00845"},{"key":"1685_CR36","doi-asserted-by":"crossref","unstructured":"Zou, H., Xu, L., Okatani, T.: Geometry enhanced reference-based image super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6123\u20136132 (2023)","DOI":"10.1109\/CVPRW59228.2023.00652"},{"key":"1685_CR37","doi-asserted-by":"crossref","unstructured":"Cao, J., Liang, J., Zhang, K., Li, Y., Zhang, Y., Wang, W., Gool, L.V.: Reference-based image super-resolution with deformable attention transformer. In: European Conference on Computer Vision. Springer, pp. 325\u2013342 (2022)","DOI":"10.1007\/978-3-031-19797-0_19"},{"key":"1685_CR38","unstructured":"Aslahishahri, M., Ubbens, J., Stavness, I.: HiTSR: A hierarchical transformer for reference-based super-resolution. arXiv preprint arXiv:2408.16959 (2024)"},{"key":"1685_CR39","doi-asserted-by":"crossref","unstructured":"Dong, R., Yuan, S., Luo, B., Chen, M., Zhang, J., Zhang, L., Li, W., Zheng, J., Fu, H.: Building bridges across spatial and temporal resolutions: Reference-based super-resolution via change priors and conditional diffusion model. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 27684\u201327694 (2024)","DOI":"10.1109\/CVPR52733.2024.02614"},{"key":"1685_CR40","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/LGRS.2023.3336680","volume":"21","author":"J Min","year":"2023","unstructured":"Min, J., Lee, Y., Kim, D., Yoo, J.: Bridging the domain gap: A simple domain matching method for reference-based image super-resolution in remote sensing. IEEE Geosci. Remote Sens. Lett. 21, 1\u20135 (2023)","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"1685_CR41","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2024.3428600","author":"H Zheng","year":"2024","unstructured":"Zheng, H., Heng, H., Yan, Z., Zeng, K., Fang, J., Qiang, B.: A generic multi-correspondence matching framework for reference-based image super-resolution. IEEE Trans. Instr. Measur. (2024). https:\/\/doi.org\/10.1109\/TIM.2024.3428600","journal-title":"IEEE Trans. Instr. Measur."},{"key":"1685_CR42","first-page":"1","volume":"62","author":"Z Tu","year":"2024","unstructured":"Tu, Z., Yang, X., He, X., Yan, J., Xu, T.: RGTGAN: Reference-based gradient-assisted texture-enhancement GAN for remote sensing super-resolution. IEEE Trans. Geosci. Remote Sens. 62, 1\u201321 (2024)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"1685_CR43","unstructured":"Lei, P., Fang, F., Zhang, G.: Flow guidance deformable compensation network for video frame interpolation. arXiv preprint arXiv:2211.12117 (2022)"},{"key":"1685_CR44","unstructured":"Li, C., Wu, G., Sun, Y., Tao, X., Tang, C.-K., Tai, Y.-W.: H-vfi: Hierarchical frame interpolation for videos with large motions. arXiv preprint arXiv:2211.11309 (2022)"},{"key":"1685_CR45","doi-asserted-by":"crossref","unstructured":"Jin, X., Wu, L., Chen, J., Chen, Y., Koo, J., Hahm, C.-h.: A unified pyramid recurrent network for video frame interpolation. arXiv preprint arXiv:2211.03456 (2022)","DOI":"10.1109\/CVPR52729.2023.00158"},{"key":"1685_CR46","doi-asserted-by":"crossref","unstructured":"Lin, X., Zhao, L., Liu, X., Chen, J.: Mvfi-net: Motion-aware video frame interpolation network. In: Proceedings of the Asian Conference on Computer Vision, pp. 3690\u20133706 (2022)","DOI":"10.1007\/978-3-031-26313-2_21"},{"issue":"12","key":"1685_CR47","doi-asserted-by":"publisher","first-page":"2980","DOI":"10.1007\/s11263-022-01683-9","volume":"130","author":"Y Li","year":"2022","unstructured":"Li, Y., Zhu, Y., Li, R., Wang, X., Luo, Y., Shan, Y.: Hybrid warping fusion for video frame interpolation. Int. J. Comput. Vision 130(12), 2980\u20132993 (2022)","journal-title":"Int. J. Comput. Vision"},{"key":"1685_CR48","doi-asserted-by":"crossref","unstructured":"Hu, M., Jiang, K., Liao, L., Nie, Z., Xiao, J., Wang, Z.: Progressive spatial-temporal collaborative network for video frame interpolation. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 2145\u20132153 (2022)","DOI":"10.1145\/3503161.3547875"},{"key":"1685_CR49","doi-asserted-by":"crossref","unstructured":"Plack, M., Briedis, K.M., Djelouah, A., Hullin, M.B., Gross, M., Schroers, C.: Frame interpolation transformer and uncertainty guidance. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9811\u20139821 (2023)","DOI":"10.1109\/CVPR52729.2023.00946"},{"key":"1685_CR50","doi-asserted-by":"crossref","unstructured":"Zhou, C., Liu, J., Tang, J., Wu, G.: Video frame interpolation with densely queried bilateral correlation. arXiv preprint arXiv:2304.13596 (2023)","DOI":"10.24963\/ijcai.2023\/198"},{"key":"1685_CR51","doi-asserted-by":"crossref","unstructured":"Li, Z., Zhu, Z.-L., Han, L.-H., Hou, Q., Guo, C.-L., Cheng, M.-M.: Amt: All-pairs multi-field transforms for efficient frame interpolation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9801\u20139810 (2023)","DOI":"10.1109\/CVPR52729.2023.00945"},{"key":"1685_CR52","doi-asserted-by":"crossref","unstructured":"Liu, C., Yang, H., Fu, J., Qian, X.: Ttvfi: Learning trajectory-aware transformer for video frame interpolation. arXiv preprint arXiv:2207.09048 (2022)","DOI":"10.1109\/TIP.2023.3302990"},{"issue":"5","key":"1685_CR53","doi-asserted-by":"publisher","first-page":"2116","DOI":"10.1109\/TCSVT.2022.3222875","volume":"33","author":"Y Zhang","year":"2022","unstructured":"Zhang, Y., Wang, H., Zhu, H., Chen, Z.: Optical flow reusing for high-efficiency space-time video super resolution. IEEE Trans. Circ. Syst. Video Technol. 33(5), 2116\u20132128 (2022)","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"key":"1685_CR54","doi-asserted-by":"publisher","first-page":"6773","DOI":"10.1109\/TIP.2022.3215911","volume":"31","author":"Y Luo","year":"2022","unstructured":"Luo, Y., Pan, J., Tang, J.: Bi-directional pseudo-three-dimensional network for video frame interpolation. IEEE Trans. Image Process. 31, 6773\u20136788 (2022)","journal-title":"IEEE Trans. Image Process."},{"issue":"9","key":"1685_CR55","doi-asserted-by":"publisher","first-page":"10535","DOI":"10.1007\/s10489-022-03787-8","volume":"53","author":"H-D Li","year":"2022","unstructured":"Li, H.-D., Yin, H., Liu, Z.-H., Huang, H.: Enhanced spatial-temporal freedom for video frame interpolation. Appl. Intell. 53(9), 10535\u201310547 (2022)","journal-title":"Appl. Intell."},{"key":"1685_CR56","doi-asserted-by":"crossref","unstructured":"Huang, Z., Zhang, T., Heng, W., Shi, B., Zhou, S.: Real-time intermediate flow estimation for video frame interpolation. In: European Conference on Computer Vision, pp. 624\u2013642 (2022). Springer","DOI":"10.1007\/978-3-031-19781-9_36"},{"key":"1685_CR57","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation. In: International Conference on Medical Image Computing and Computer-assisted Intervention. Springer, pp. 234\u2013241 (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"1685_CR58","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"}],"container-title":["Machine Vision and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-025-01685-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00138-025-01685-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-025-01685-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,9]],"date-time":"2025-05-09T14:30:48Z","timestamp":1746801048000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00138-025-01685-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,23]]},"references-count":58,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,5]]}},"alternative-id":["1685"],"URL":"https:\/\/doi.org\/10.1007\/s00138-025-01685-w","relation":{},"ISSN":["0932-8092","1432-1769"],"issn-type":[{"value":"0932-8092","type":"print"},{"value":"1432-1769","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,4,23]]},"assertion":[{"value":"11 October 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 March 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 March 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 April 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"69"}}