{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T11:31:26Z","timestamp":1768908686802,"version":"3.49.0"},"reference-count":35,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,9,16]],"date-time":"2024-09-16T00:00:00Z","timestamp":1726444800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,9,16]],"date-time":"2024-09-16T00:00:00Z","timestamp":1726444800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61573182"],"award-info":[{"award-number":["61573182"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62073164"],"award-info":[{"award-number":["62073164"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["NS2022041"],"award-info":[{"award-number":["NS2022041"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,3]]},"DOI":"10.1007\/s00371-024-03621-x","type":"journal-article","created":{"date-parts":[[2024,9,16]],"date-time":"2024-09-16T08:04:08Z","timestamp":1726473848000},"page":"3581-3593","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["FDDCC-VSR: a lightweight video super-resolution network based on deformable 3D convolution and cheap convolution"],"prefix":"10.1007","volume":"41","author":[{"given":"Xiaohu","family":"Wang","sequence":"first","affiliation":[]},{"given":"Xin","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Hengrui","family":"Li","sequence":"additional","affiliation":[]},{"given":"Tao","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,16]]},"reference":[{"issue":"1","key":"3621_CR1","doi-asserted-by":"publisher","first-page":"43","DOI":"10.1093\/comjnl\/bxm075","volume":"52","author":"H Greenspan","year":"2009","unstructured":"Greenspan, H.: Super-Resolution in Medical Imaging[J]. Comput. J. 52(1), 43\u201363 (2009)","journal-title":"Comput. J."},{"key":"3621_CR2","unstructured":"Lillesand, T., Kiefer, R.W. and Chipman, J.: Remote sensing and image interpretation, (2014)"},{"key":"3621_CR3","unstructured":"Lobanov, A. P.: Resolution limits in astronomical images, arXiv preprint astro-ph\/0503225 (2005)"},{"key":"3621_CR4","unstructured":"Caballero, J., Ledig, C., Aitken, A., et al.: Real-time video super-resolution with spatio-temporal networks and motion compensation: U.S. Patent 10,701,394[P]. (2020)"},{"issue":"22","key":"3621_CR5","doi-asserted-by":"publisher","first-page":"2873","DOI":"10.3390\/math9222873","volume":"9","author":"A Khan","year":"2021","unstructured":"Khan, A., Sargano, A.B., Habib, Z.: DSTnet: deformable spatio-temporal convolutional residual network for video super-resolution[J]. Mathematics 9(22), 2873 (2021)","journal-title":"Mathematics"},{"key":"3621_CR6","doi-asserted-by":"crossref","unstructured":"Song, C., Lin, Y., Guo, S., et al.: Spatial-temporal synchronous graph convolutional networks: a new framework for spatial-temporal network data forecasting[C]\/\/Proceedings of the AAAI Conference on Artificial Intelligence. 34(01): 914\u2013921 (2020)","DOI":"10.1609\/aaai.v34i01.5438"},{"key":"3621_CR7","doi-asserted-by":"crossref","unstructured":"Yang, X., Li, H., Li, X.: Lightweight image super-resolution with feature cheap convolution and attention mechanism[J]. Cluster Computing, 1\u201316 (2022)","DOI":"10.1007\/s10586-022-03631-1"},{"issue":"5","key":"3621_CR8","doi-asserted-by":"publisher","first-page":"7063","DOI":"10.1007\/s11042-020-09958-4","volume":"80","author":"X Yang","year":"2021","unstructured":"Yang, X., Zhang, Y., Guo, Y., et al.: An image super-resolution deep learning network based on multi-level feature extraction module[J]. Multimedia Tools and Applications 80(5), 7063\u20137075 (2021)","journal-title":"Multimedia Tools and Applications"},{"key":"3621_CR9","unstructured":"Xie, Z., Zhang, W., Sheng, B., et al.: BaGFN: broad attentive graph fusion network for high-order feature interactions[J]. IEEE Trans. Neural Netw. Learn. Syst. (2021)"},{"key":"3621_CR10","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition[J]. arXiv preprint arXiv:1409.1556, (2014)"},{"key":"3621_CR11","unstructured":"Lin, X., Sun, S., Huang, W., et al.: EAPT: efficient attention pyramid transformer for image processing[J]. IEEE Transactions on Multimedia, (2021)"},{"key":"3621_CR12","doi-asserted-by":"crossref","unstructured":"Jiang, N., Sheng, B., Li, P., et al.: PhotoHelper: portrait photographing guidance via deep feature retrieval and fusion[J]. IEEE Transactions on Multimedia, (2022)","DOI":"10.1109\/TMM.2022.3144890"},{"key":"3621_CR13","doi-asserted-by":"crossref","unstructured":"Liu, S., Huang, D.: Receptive field block net for accurate and fast object detection[C]\/\/Proceedings of the European conference on computer vision (ECCV). 385\u2013400, (2018)","DOI":"10.1007\/978-3-030-01252-6_24"},{"key":"3621_CR14","doi-asserted-by":"publisher","first-page":"1500","DOI":"10.1109\/LSP.2020.3013518","volume":"27","author":"X Ying","year":"2020","unstructured":"Ying, X., Wang, L., Wang, Y., et al.: Deformable 3d convolution for video super-resolution[J]. IEEE Signal Process. Lett. 27, 1500\u20131504 (2020)","journal-title":"IEEE Signal Process. Lett."},{"key":"3621_CR15","doi-asserted-by":"crossref","unstructured":"Dai, J., Qi, H., Xiong, Y., et al.: Deformable convolutional networks[C]\/\/Proceedings of the IEEE international conference on computer vision. 764\u2013773 (2017)","DOI":"10.1109\/ICCV.2017.89"},{"key":"3621_CR16","doi-asserted-by":"crossref","unstructured":"Wang, R., Shivanna, R., Cheng, D., et al. Dcn v2: Improved deep & cross network and practical lessons for web-scale learning to rank systems[C]\/\/Proceedings of the Web Conference 2021.1785\u20131797 (2021)","DOI":"10.1145\/3442381.3450078"},{"key":"3621_CR17","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Zhang, M., Lu, F.: Optical flow in the dark[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 6749\u20136757 (2020)","DOI":"10.1109\/CVPR42600.2020.00678"},{"key":"3621_CR18","doi-asserted-by":"crossref","unstructured":"Han, K., Wang, Y., Tian, Q., et al.: Ghostnet: More features from cheap operations[C]\/\/Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. 1580\u20131589 (2020)","DOI":"10.1109\/CVPR42600.2020.00165"},{"key":"3621_CR19","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network[J]. arXiv preprint arXiv:1503.02531, 2(7) (2015)"},{"key":"3621_CR20","unstructured":"Gui, S., Wang, H., Yu, C., et al.: Adversarially trained model compression: when robustness meets efficiency[J]. arXiv preprint arXiv:1902.03538, (2019)"},{"issue":"8","key":"3621_CR21","doi-asserted-by":"publisher","first-page":"1106","DOI":"10.1007\/s11263-018-01144-2","volume":"127","author":"T Xue","year":"2019","unstructured":"Xue, T., Chen, B., Wu, J., et al.: Video enhancement with task-oriented flow[J]. Int. J. Comput. Vision 127(8), 1106\u20131125 (2019)","journal-title":"Int. J. Comput. Vision"},{"key":"3621_CR22","doi-asserted-by":"crossref","unstructured":"Lim, B., Son, S., Kim, H., et al.: Enhanced deep residual networks for single image super-resolution[C]\/\/Proceedings of the IEEE conference on computer vision and pattern recognition workshops. 136\u2013144 (2017)","DOI":"10.1109\/CVPRW.2017.151"},{"key":"3621_CR23","doi-asserted-by":"crossref","unstructured":"Tao, X., Gao, H., Liao, R., et al.: Detail-revealing deep video super-resolution[C]\/\/Proceedings of the IEEE International Conference on Computer Vision. 4472\u20134480 (2017)","DOI":"10.1109\/ICCV.2017.479"},{"key":"3621_CR24","doi-asserted-by":"crossref","unstructured":"Haris, M., Shakhnarovich, G., Ukita, N.: Deep back-projection networks for super-resolution[C]\/\/Proceedings of the IEEE conference on computer vision and pattern recognition. 1664\u20131673 (2018)","DOI":"10.1109\/CVPR.2018.00179"},{"key":"3621_CR25","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Li, K., Li, K., et al. Image super-resolution using very deep residual channel attention networks[C]\/\/Proceedings of the European conference on computer vision (ECCV). 286\u2013301 (2018)","DOI":"10.1007\/978-3-030-01234-2_18"},{"key":"3621_CR26","doi-asserted-by":"crossref","unstructured":"Yi, P., Wang, Z., Jiang, K., et al.: Progressive fusion video super-resolution network via exploiting non-local spatio-temporal correlations[C]\/\/Proceedings of the IEEE\/CVF international conference on computer vision. 3106\u20133115 (2019)","DOI":"10.1109\/ICCV.2019.00320"},{"key":"3621_CR27","doi-asserted-by":"crossref","unstructured":"Wang, X., Chan, K.C.K., Yu, K., et al.: Edvr: Video restoration with enhanced deformable convolutional networks[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops. 0\u20130 (2019)","DOI":"10.1109\/CVPRW.2019.00247"},{"key":"3621_CR28","doi-asserted-by":"publisher","first-page":"4323","DOI":"10.1109\/TIP.2020.2967596","volume":"29","author":"L Wang","year":"2020","unstructured":"Wang, L., Guo, Y., Liu, L., et al.: Deep video super-resolution using HR optical flow estimation[J]. IEEE Trans. Image Process. 29, 4323\u20134336 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"3621_CR29","doi-asserted-by":"crossref","unstructured":"Tian, Y., Zhang, Y., Fu, Y., et al.: Tdan: Temporally-deformable alignment network for video super-resolution[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 3360\u20133369 (2020)","DOI":"10.1109\/CVPR42600.2020.00342"},{"key":"3621_CR30","doi-asserted-by":"crossref","unstructured":"Yi, P., Wang, Z., Jiang, K., et al.: Omniscient video super-resolution[C]\/\/Proceedings of the IEEE\/CVF International Conference on Computer Vision. 4429\u20134438 (2021)","DOI":"10.1109\/ICCV48922.2021.00439"},{"key":"3621_CR31","doi-asserted-by":"crossref","unstructured":"Xiang, X., Tian, Y., Zhang, Y., et al.: Zooming slow-mo: Fast and accurate one-stage space-time video super-resolution[C]\/\/Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. 3370\u20133379 (2020)","DOI":"10.1109\/CVPR42600.2020.00343"},{"key":"3621_CR32","doi-asserted-by":"crossref","unstructured":"Geng, Z., Liang, L., Ding, T., et al.: RSTT: Real-time Spatial Temporal Transformer for Space-Time Video Super-Resolution[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 17441\u201317451 (2022)","DOI":"10.1109\/CVPR52688.2022.01692"},{"key":"3621_CR33","unstructured":"Wang, H., Xiang, X., Tian, Y., et al.: STDAN: Deformable Attention Network for Space-Time Video Super-Resolution[J]. arXiv preprint arXiv:2203.06841, (2022)"},{"key":"3621_CR34","doi-asserted-by":"crossref","unstructured":"Xu, G., Xu, J., Li, Z., et al.: Temporal modulation network for controllable space-time video super-resolution[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 6388\u20136397 (2021)","DOI":"10.1109\/CVPR46437.2021.00632"},{"key":"3621_CR35","unstructured":"Crowley, E.J., Gray, G., Storkey, A.J.: Moonshine: Distilling with cheap convolutions[J]. Advances in Neural Information Processing Systems, 31 (2018)"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03621-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-024-03621-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03621-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,10]],"date-time":"2025-03-10T09:11:36Z","timestamp":1741597896000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-024-03621-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,16]]},"references-count":35,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,3]]}},"alternative-id":["3621"],"URL":"https:\/\/doi.org\/10.1007\/s00371-024-03621-x","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,9,16]]},"assertion":[{"value":"27 August 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 September 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Xiaohu Wang declares that he has no conflict of interest. Xin Yang declares that he has no conflict of interest. Hengrui Li declares that she has no conflict of interest. Tao Li declares that he has no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}