{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,8]],"date-time":"2026-02-08T19:12:52Z","timestamp":1770577972721,"version":"3.49.0"},"reference-count":54,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T00:00:00Z","timestamp":1722470400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T00:00:00Z","timestamp":1722470400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T00:00:00Z","timestamp":1722470400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T00:00:00Z","timestamp":1722470400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T00:00:00Z","timestamp":1722470400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T00:00:00Z","timestamp":1722470400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T00:00:00Z","timestamp":1722470400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Journal of Visual Communication and Image Representation"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1016\/j.jvcir.2024.104226","type":"journal-article","created":{"date-parts":[[2024,7,9]],"date-time":"2024-07-09T23:42:54Z","timestamp":1720568574000},"page":"104226","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":3,"special_numbering":"C","title":["EMCFN: Edge-based Multi-scale Cross Fusion Network for video frame interpolation"],"prefix":"10.1016","volume":"103","author":[{"given":"Shaowen","family":"Wang","sequence":"first","affiliation":[]},{"given":"Xiaohui","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Zhiquan","family":"Feng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6157-2051","authenticated-orcid":false,"given":"Jiande","family":"Sun","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3160-9233","authenticated-orcid":false,"given":"Ju","family":"Liu","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.jvcir.2024.104226_b1","doi-asserted-by":"crossref","unstructured":"L. Siyao, S. Zhao, W. Yu, W. Sun, D. Metaxas, C.C. Loy, Z. Liu, Deep Animation Video Interpolation in the Wild, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2021, pp. 6587\u20136595.","DOI":"10.1109\/CVPR46437.2021.00652"},{"key":"10.1016\/j.jvcir.2024.104226_b2","doi-asserted-by":"crossref","unstructured":"T. Ding, L. Liang, Z. Zhu, I. Zharkov, CDFI: Compression-Driven Network Design for Frame Interpolation, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2021, pp. 8001\u20138011.","DOI":"10.1109\/CVPR46437.2021.00791"},{"key":"10.1016\/j.jvcir.2024.104226_b3","doi-asserted-by":"crossref","unstructured":"H. Lee, T. Kim, T.-y. Chung, D. Pak, Y. Ban, S. Lee, AdaCoF: Adaptive Collaboration of Flows for Video Frame Interpolation, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2020, pp. 5316\u20135325.","DOI":"10.1109\/CVPR42600.2020.00536"},{"key":"10.1016\/j.jvcir.2024.104226_b4","doi-asserted-by":"crossref","unstructured":"J. Flynn, I. Neulander, J. Philbin, N. Snavely, DeepStereo: Learning to Predict New Views From the World\u2019s Imagery, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2016, pp. 5515\u20135524.","DOI":"10.1109\/CVPR.2016.595"},{"key":"10.1016\/j.jvcir.2024.104226_b5","doi-asserted-by":"crossref","unstructured":"T. Peleg, P. Szekely, D. Sabo, O. Sendik, IM-Net for High Resolution Video Frame Interpolation, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2019, pp. 2398\u20132407.","DOI":"10.1109\/CVPR.2019.00250"},{"key":"10.1016\/j.jvcir.2024.104226_b6","doi-asserted-by":"crossref","unstructured":"L. Kong, B. Jiang, D. Luo, W. Chu, X. Huang, Y. Tai, C. Wang, J. Yang, IFRNet: Intermediate Feature Refine Network for Efficient Frame Interpolation, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2022, pp. 1969\u20131978.","DOI":"10.1109\/CVPR52688.2022.00201"},{"key":"10.1016\/j.jvcir.2024.104226_b7","series-title":"Proc. IEEE\/CVF Int. Conf. Comput. Vis. Workshop","first-page":"3427","article-title":"Quadratic Video Interpolation for VTSR Challenge","author":"Li","year":"2019"},{"key":"10.1016\/j.jvcir.2024.104226_b8","doi-asserted-by":"crossref","unstructured":"E. Ilg, N. Mayer, T. Saikia, M. Keuper, A. Dosovitskiy, T. Brox, FlowNet 2.0: Evolution of Optical Flow Estimation with Deep Networks, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2017, pp. 2462\u20132470.","DOI":"10.1109\/CVPR.2017.179"},{"key":"10.1016\/j.jvcir.2024.104226_b9","doi-asserted-by":"crossref","unstructured":"D. Sun, X. Yang, M.-Y. Liu, J. Kautz, PWC-Net: CNNs for Optical Flow Using Pyramid, Warping, and Cost Volume, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2018, pp. 8934\u20138943.","DOI":"10.1109\/CVPR.2018.00931"},{"key":"10.1016\/j.jvcir.2024.104226_b10","doi-asserted-by":"crossref","unstructured":"H. Jiang, D. Sun, V. Jampani, M.-H. Yang, E. Learned-Miller, J. Kautz, Super SloMo: High Quality Estimation of Multiple Intermediate Frames for Video Interpolation, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2018, pp. 9000\u20139008.","DOI":"10.1109\/CVPR.2018.00938"},{"key":"10.1016\/j.jvcir.2024.104226_b11","doi-asserted-by":"crossref","unstructured":"S. Niklaus, L. Mai, F. Liu, Video Frame Interpolation via Adaptive Separable Convolution, in: Proc. IEEE Int. Conf. Comput. Vis., ICCV, 2017, pp. 261\u2013270.","DOI":"10.1109\/ICCV.2017.37"},{"key":"10.1016\/j.jvcir.2024.104226_b12","series-title":"Proc. IEEE Conf. Acoust. Speech Signal Process.","first-page":"1958","article-title":"Video Frame Interpolation Via Exceptional Motion-Aware Synthesis","author":"Park","year":"2020"},{"key":"10.1016\/j.jvcir.2024.104226_b13","doi-asserted-by":"crossref","unstructured":"Z. Liu, R.A. Yeh, X. Tang, Y. Liu, A. Agarwala, Video Frame Synthesis using Deep Voxel Flow, in: Proc. IEEE Int. Conf. Comput. Vis., ICCV, 2017, pp. 4463\u20134471.","DOI":"10.1109\/ICCV.2017.478"},{"key":"10.1016\/j.jvcir.2024.104226_b14","doi-asserted-by":"crossref","unstructured":"W. Bao, W.-S. Lai, C. Ma, X. Zhang, Z. Gao, M.-H. Yang, Depth-Aware Video Frame Interpolation, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2019, pp. 3703\u20133712.","DOI":"10.1109\/CVPR.2019.00382"},{"issue":"3","key":"10.1016\/j.jvcir.2024.104226_b15","doi-asserted-by":"crossref","first-page":"933","DOI":"10.1109\/TPAMI.2019.2941941","article-title":"MEMC-Net: Motion Estimation and Motion Compensation Driven Neural Network for Video Interpolation and Enhancement","volume":"43","author":"Bao","year":"2019","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.jvcir.2024.104226_b16","doi-asserted-by":"crossref","unstructured":"S. Niklaus, L. Mai, F. Liu, Video Frame Interpolation via Adaptive Convolution, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2017, pp. 670\u2013679.","DOI":"10.1109\/CVPR.2017.244"},{"issue":"5","key":"10.1016\/j.jvcir.2024.104226_b17","doi-asserted-by":"crossref","first-page":"1953","DOI":"10.1109\/TCSVT.2020.3011197","article-title":"Compression Priors Assisted Convolutional Neural Network for Fractional Interpolation","volume":"31","author":"Zhang","year":"2020","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"3","key":"10.1016\/j.jvcir.2024.104226_b18","doi-asserted-by":"crossref","first-page":"840","DOI":"10.1109\/TCSVT.2018.2816932","article-title":"Convolutional Neural Network-Based Fractional-Pixel Motion Compensation","volume":"29","author":"Yan","year":"2018","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.jvcir.2024.104226_b19","series-title":"International Conference Medical Image Computing and Computer-Assisted Intervention","first-page":"234","article-title":"U-Net: Convolutional Networks for Biomedical Image Segmentation","author":"Ronneberger","year":"2015"},{"key":"10.1016\/j.jvcir.2024.104226_b20","doi-asserted-by":"crossref","unstructured":"S. Woo, J. Park, J.-Y. Lee, I.S. Kweon, CBAM: Convolutional Block Attention Module, in: Proc. Eur. Conf. Comput. Vis., ECCV, 2018, pp. 3\u201319.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"10.1016\/j.jvcir.2024.104226_b21","doi-asserted-by":"crossref","unstructured":"D. Fourure, R. Emonet, E. Fromont, D. Muselet, A. Tremeau, C. Wolf, Residual Conv-Deconv Grid Network for Semantic Segmentation, in: Proc. British Mach. Vis. Conf., 2017.","DOI":"10.5244\/C.31.181"},{"key":"10.1016\/j.jvcir.2024.104226_b22","series-title":"Proc. Eur. Conf. Comput. Vis.","first-page":"250","article-title":"FILM: Frame Interpolation for Large Motion","author":"Reda","year":"2022"},{"key":"10.1016\/j.jvcir.2024.104226_b23","article-title":"Optical Flow Reusing for High-Efficiency Space-Time Video Super Resolution","author":"Zhang","year":"2022","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.jvcir.2024.104226_b24","doi-asserted-by":"crossref","unstructured":"D. Danier, F. Zhang, D. Bull, ST-MFNet: A Spatio-Temporal Multi-Flow Network for Frame Interpolation, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2022, pp. 3521\u20133531.","DOI":"10.1109\/CVPR52688.2022.00351"},{"issue":"1","key":"10.1016\/j.jvcir.2024.104226_b25","doi-asserted-by":"crossref","first-page":"359","DOI":"10.1109\/TCSVT.2021.3061153","article-title":"Spatiotemporal Generative Adversarial Network-Based Dynamic Texture Synthesis for Surveillance Video Coding","volume":"32","author":"Yang","year":"2021","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"11","key":"10.1016\/j.jvcir.2024.104226_b26","doi-asserted-by":"crossref","first-page":"3968","DOI":"10.1109\/TCSVT.2019.2939143","article-title":"A Multi-Scale Position Feature Transform Network for Video Frame Interpolation","volume":"30","author":"Cheng","year":"2019","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.jvcir.2024.104226_b27","series-title":"Proc. Eur. Conf. Comput. Vis.","first-page":"109","article-title":"BMBC: Bilateral Motion Estimation with Bilateral Cost Volume for Video Interpolation","author":"Park","year":"2020"},{"issue":"2","key":"10.1016\/j.jvcir.2024.104226_b28","first-page":"1","article-title":"L2BEC2: Local Lightweight Bidirectional Encoding and Channel Attention Cascade for Video Frame Interpolation","volume":"19","author":"Zhang","year":"2023","journal-title":"ACM Trans. Multimed. Comput. Commun. Appl."},{"issue":"10","key":"10.1016\/j.jvcir.2024.104226_b29","doi-asserted-by":"crossref","first-page":"7029","DOI":"10.1109\/TPAMI.2021.3100714","article-title":"Multiple Video Frame Interpolation via Enhanced Deformable Separable Convolution","volume":"44","author":"Cheng","year":"2021","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.jvcir.2024.104226_b30","article-title":"Edge-Aware Network for Flow-Based Video Frame Interpolation","author":"Zhao","year":"2022","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"issue":"12","key":"10.1016\/j.jvcir.2024.104226_b31","doi-asserted-by":"crossref","first-page":"2980","DOI":"10.1007\/s11263-022-01683-9","article-title":"Hybrid Warping Fusion for Video Frame Interpolation","volume":"130","author":"Li","year":"2022","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.jvcir.2024.104226_b32","doi-asserted-by":"crossref","unstructured":"Y.-L. Liu, Y.-T. Liao, Y.-Y. Lin, Y.-Y. Chuang, Deep Video Frame Interpolation Using Cyclic Frame Generation, in: Proc. AAAI Conf. Artif. Intell., Vol. 33, 2019, pp. 8794\u20138802.","DOI":"10.1609\/aaai.v33i01.33018794"},{"key":"10.1016\/j.jvcir.2024.104226_b33","doi-asserted-by":"crossref","first-page":"3726","DOI":"10.1109\/TIP.2022.3175432","article-title":"DO-Conv: Depthwise Over-Parameterized Convolutional Layer","volume":"31","author":"Cao","year":"2022","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.jvcir.2024.104226_b34","doi-asserted-by":"crossref","DOI":"10.1016\/j.jvcir.2022.103735","article-title":"Edge-aware object pixel-level representation tracking","volume":"90","author":"Jing","year":"2023","journal-title":"J. Vis. Commun. Image Represent."},{"key":"10.1016\/j.jvcir.2024.104226_b35","article-title":"Flow Guidance Deformable Compensation Network for Video Frame Interpolation","author":"Lei","year":"2023","journal-title":"IEEE Trans. Multimed."},{"issue":"6","key":"10.1016\/j.jvcir.2024.104226_b36","doi-asserted-by":"crossref","first-page":"3390","DOI":"10.1109\/TCSVT.2021.3110796","article-title":"Capturing Small, Fast-Moving Objects: Frame Interpolation via Recurrent Motion Enhancement","volume":"32","author":"Hu","year":"2021","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"3","key":"10.1016\/j.jvcir.2024.104226_b37","doi-asserted-by":"crossref","first-page":"1178","DOI":"10.1109\/TCSVT.2020.2995243","article-title":"Deep Network-Based Frame Extrapolation With Reference Frame Alignment","volume":"31","author":"Huo","year":"2020","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.jvcir.2024.104226_b38","doi-asserted-by":"crossref","unstructured":"K. He, X. Zhang, S. Ren, J. Sun, Delving Deep into Rectifiers: Surpassing Human-Level Performance on Imagenet Classification, in: Proc. IEEE Int. Conf. Comput. Vis., ICCV, 2015, pp. 1026\u20131034.","DOI":"10.1109\/ICCV.2015.123"},{"key":"10.1016\/j.jvcir.2024.104226_b39","doi-asserted-by":"crossref","unstructured":"M. Liang, B. Yang, Y. Chen, R. Hu, R. Urtasun, Multi-Task Multi-Sensor Fusion for 3D Object Detection, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2019, pp. 7345\u20137353.","DOI":"10.1109\/CVPR.2019.00752"},{"key":"10.1016\/j.jvcir.2024.104226_b40","series-title":"Proc. Eur. Conf. Comput. Vis.","first-page":"434","article-title":"Learning Image Matching by Simply Watching Video","author":"Long","year":"2016"},{"key":"10.1016\/j.jvcir.2024.104226_b41","article-title":"PyTorch: An Imperative Style, High-Performance Deep Learning Library","volume":"32","author":"Paszke","year":"2019","journal-title":"NeurIPS"},{"key":"10.1016\/j.jvcir.2024.104226_b42","series-title":"Adam: A Method for Stochastic Optimization","author":"Kingma","year":"2014"},{"key":"10.1016\/j.jvcir.2024.104226_b43","doi-asserted-by":"crossref","first-page":"1106","DOI":"10.1007\/s11263-018-01144-2","article-title":"Video Enhancement with Task-Oriented Flow","volume":"127","author":"Xue","year":"2019","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.jvcir.2024.104226_b44","series-title":"Proc. Int. Conf. Pattern Recognit.","first-page":"2366","article-title":"Image Quality Metrics: PSNR vs. SSIM","author":"Hore","year":"2010"},{"issue":"4","key":"10.1016\/j.jvcir.2024.104226_b45","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image quality assessment: from error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.jvcir.2024.104226_b46","doi-asserted-by":"crossref","unstructured":"Z. Liu, R.A. Yeh, X. Tang, Y. Liu, A. Agarwala, Video Frame Synthesis Using Deep Voxel Flow, in: Proc. IEEE Int. Conf. Comput. Vis., ICCV, 2017, pp. 4463\u20134471.","DOI":"10.1109\/ICCV.2017.478"},{"key":"10.1016\/j.jvcir.2024.104226_b47","doi-asserted-by":"crossref","unstructured":"F. Perazzi, J. Pont-Tuset, B. McWilliams, L. Van Gool, M. Gross, A. Sorkine-Hornung, A Benchmark Dataset and Evaluation Methodology for Video Object Segmentation, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2016, pp. 724\u2013732.","DOI":"10.1109\/CVPR.2016.85"},{"key":"10.1016\/j.jvcir.2024.104226_b48","series-title":"Proc. IEEE Int. Conf. Acoust. Speech Signal Process.","first-page":"2613","article-title":"Video Frame Interpolation Via Residue Refinement","author":"Li","year":"2020"},{"key":"10.1016\/j.jvcir.2024.104226_b49","doi-asserted-by":"crossref","unstructured":"X. Jin, L. Wu, J. Chen, Y. Chen, J. Koo, C.-h. Hahm, A Unified Pyramid Recurrent Network for Video Frame Interpolation, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2023, pp. 1578\u20131587.","DOI":"10.1109\/CVPR52729.2023.00158"},{"key":"10.1016\/j.jvcir.2024.104226_b50","doi-asserted-by":"crossref","unstructured":"D. Zhang, P. Huang, X. Ding, F. Li, G. Yang, Video Frame Interpolation via Multi-scale Expandable Deformable Convolution, in: Proceedings of the 2023 ACM Workshop on Information Hiding and Multimedia Security, 2023, pp. 19\u201328.","DOI":"10.1145\/3577163.3595098"},{"key":"10.1016\/j.jvcir.2024.104226_b51","doi-asserted-by":"crossref","unstructured":"S. Niklaus, F. Liu, Context-Aware Synthesis for Video Frame Interpolation, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2018, pp. 1701\u20131710.","DOI":"10.1109\/CVPR.2018.00183"},{"key":"10.1016\/j.jvcir.2024.104226_b52","doi-asserted-by":"crossref","unstructured":"L. Lu, R. Wu, H. Lin, J. Lu, J. Jia, Video Frame Interpolation with Transformer, in: Proc. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit., CVPR, 2022, pp. 3532\u20133542.","DOI":"10.1109\/CVPR52688.2022.00352"},{"key":"10.1016\/j.jvcir.2024.104226_b53","doi-asserted-by":"crossref","unstructured":"C. Bucilu\u01ce, R. Caruana, A. Niculescu-Mizil, Model compression, in: Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2006, pp. 535\u2013541.","DOI":"10.1145\/1150402.1150464"},{"key":"10.1016\/j.jvcir.2024.104226_b54","series-title":"Distilling the Knowledge in a Neural Network","author":"Hinton","year":"2015"}],"container-title":["Journal of Visual Communication and Image Representation"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1047320324001822?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1047320324001822?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2024,9,5]],"date-time":"2024-09-05T00:19:21Z","timestamp":1725495561000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1047320324001822"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8]]},"references-count":54,"alternative-id":["S1047320324001822"],"URL":"https:\/\/doi.org\/10.1016\/j.jvcir.2024.104226","relation":{},"ISSN":["1047-3203"],"issn-type":[{"value":"1047-3203","type":"print"}],"subject":[],"published":{"date-parts":[[2024,8]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"EMCFN: Edge-based Multi-scale Cross Fusion Network for video frame interpolation","name":"articletitle","label":"Article Title"},{"value":"Journal of Visual Communication and Image Representation","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.jvcir.2024.104226","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2024 Elsevier Inc. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"104226"}}