{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T19:23:10Z","timestamp":1768072990232,"version":"3.49.0"},"reference-count":51,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2023,4,11]],"date-time":"2023-04-11T00:00:00Z","timestamp":1681171200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,4,11]],"date-time":"2023-04-11T00:00:00Z","timestamp":1681171200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,2]]},"DOI":"10.1007\/s00371-023-02832-y","type":"journal-article","created":{"date-parts":[[2023,4,11]],"date-time":"2023-04-11T14:03:13Z","timestamp":1681221793000},"page":"1069-1082","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Trapezoid-structured LSTM with segregated gates and bridge joints for video frame inpainting"],"prefix":"10.1007","volume":"40","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0886-2479","authenticated-orcid":false,"given":"Ting-Hui","family":"Chiang","sequence":"first","affiliation":[]},{"given":"Yun-Tang","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Jaden Chao-Ho","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Yu-Chee","family":"Tseng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,11]]},"reference":[{"issue":"7553","key":"2832_CR1","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436\u2013444 (2015)","journal-title":"Nature"},{"key":"2832_CR2","volume-title":"Deep learning","author":"I Goodfellow","year":"2016","unstructured":"Goodfellow, I., Bengio, Y., Courville, A.: Deep learning. MIT Press, Cambridge (2016)"},{"key":"2832_CR3","doi-asserted-by":"crossref","unstructured":"Wang, H., Schmid, C.: Action recognition with improved trajectories. IEEE ICCV pp. 3551\u20133558 (2013)","DOI":"10.1109\/ICCV.2013.441"},{"key":"2832_CR4","unstructured":"Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition in videos. NIPS pp. 568\u2013576 (2014)"},{"key":"2832_CR5","doi-asserted-by":"crossref","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., Paluri, M.: Learning spatiotemporal features with 3D convolutional networks. IEEE ICCV pp. 4489\u20134497 (2015)","DOI":"10.1109\/ICCV.2015.510"},{"key":"2832_CR6","doi-asserted-by":"crossref","unstructured":"Wang, L., Xiong, Y., Wang, Z., Qiao, Y., Lin, D., Tang, X., Van Gool, L.: Temporal segment networks: towards good practices for deep action recognition. Springer ECCV pp. 20\u201336 (2016)","DOI":"10.1007\/978-3-319-46484-8_2"},{"key":"2832_CR7","doi-asserted-by":"crossref","unstructured":"Chiu, S.-Y., Tseng, Y.-C., Chen, J.-J.: Low-resolution thermal sensor-guided image synthesis. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision (WACV) workshops (2023)","DOI":"10.1109\/WACVW58289.2023.00011"},{"key":"2832_CR8","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2022.3219163","author":"JY Li","year":"2022","unstructured":"Li, J.Y., Lin, J.C.H., Wu, K.R., Tseng, Y.C.: Sensepred: guiding video prediction by wearable sensors. IEEE Internet Things J (2022). https:\/\/doi.org\/10.1109\/JIOT.2022.3219163","journal-title":"IEEE Internet Things J"},{"key":"2832_CR9","doi-asserted-by":"crossref","unstructured":"Van, L., Zhang, L., Chang, C., Tong, K., Wu, K., Tseng, Y.: Things in the air: tagging wearable iot information on drone videos. Discov. Internet Things 1(1), (2021)","DOI":"10.1007\/s43926-021-00005-8"},{"key":"2832_CR10","doi-asserted-by":"crossref","unstructured":"Lin, J., Gan, C., Han, S.: TSM: Temporal shift module for efficient video understanding. IEEE ICCV pp. 7082\u20137092 (2019)","DOI":"10.1109\/ICCV.2019.00718"},{"key":"2832_CR11","doi-asserted-by":"crossref","unstructured":"Zolfaghari, M., Singh, K., Brox, T.: ECO: Efficient convolutional network for online video understanding. ECCV pp. 695\u2013712 (2018)","DOI":"10.1007\/978-3-030-01216-8_43"},{"key":"2832_CR12","doi-asserted-by":"crossref","unstructured":"Wu, C.-Y., Feichtenhofer, C., Fan, H., He, K., Krahenbuhl, P., Girshick, R.: Long-Term feature banks for detailed video understanding. IEEE CVPR pp. 284\u2013293 (2019)","DOI":"10.1109\/CVPR.2019.00037"},{"key":"2832_CR13","doi-asserted-by":"crossref","unstructured":"Chiang, T., Tseng, Y., Tseng, Y.: A multi-embedding neural model for incident video retrieval. Pattern Recognition 130, 108807 (2022) [Online]. Available: https:\/\/doi.org\/10.1016\/j.patcog.2022.108807","DOI":"10.1016\/j.patcog.2022.108807"},{"issue":"3","key":"2832_CR14","doi-asserted-by":"publisher","first-page":"463","DOI":"10.1109\/TPAMI.2007.60","volume":"29","author":"Y Wexler","year":"2007","unstructured":"Wexler, Y., Shechtman, E., Irani, M.: Space-time completion of video. IEEE Trans. Pattern Anal. Mach. Intell. 29(3), 463\u2013476 (2007)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2832_CR15","doi-asserted-by":"crossref","unstructured":"Cheung, S.C.S., Zhao, J., Venkatesh, M.V.: Efficient object-based video inpainting. IEEE ICIP pp. 705\u2013708 (2006)","DOI":"10.1109\/ICIP.2006.312432"},{"issue":"8","key":"2832_CR16","doi-asserted-by":"publisher","first-page":"2953","DOI":"10.1109\/TCSVT.2020.3034422","volume":"31","author":"C Wang","year":"2021","unstructured":"Wang, C., Chen, X., Min, S., Zha, Z.-J., Wang, J.: Structure-guided deep video inpainting. IEEE Trans. Circuits Syst. Video Technol. 31(8), 2953\u20132965 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"2832_CR17","doi-asserted-by":"crossref","unstructured":"Li, Z., Lu, C.Z., Qin, J., Guo, C.L., Cheng, M.M.: Towards an end-to-end framework for flow-guided video inpainting. IEEE CVPR pp. 17562\u201317571 (2022)","DOI":"10.1109\/CVPR52688.2022.01704"},{"key":"2832_CR18","doi-asserted-by":"crossref","unstructured":"Zou, X., Yang, L., Liu, D., Lee, Y.J.: Progressive temporal feature alignment network for video inpainting. IEEE CVPR pp. 16448\u201316457 (2021)","DOI":"10.1109\/CVPR46437.2021.01618"},{"key":"2832_CR19","doi-asserted-by":"crossref","unstructured":"Ouyang, H., Wang, T., Chen, Q.: Internal video inpainting by implicit long-range propagation. IEEE ICCV pp. 14579\u201314588 (2021)","DOI":"10.1109\/ICCV48922.2021.01431"},{"issue":"3","key":"2832_CR20","doi-asserted-by":"publisher","first-page":"818","DOI":"10.1137\/S1064827501386481","volume":"24","author":"A Borzi","year":"2003","unstructured":"Borzi, A., Ito, K., Kunisch, K.: Optimal control formulation for determining optical flow. SIAM J. Sci. Comput. 24(3), 818\u2013847 (2003)","journal-title":"SIAM J. Sci. Comput."},{"issue":"3","key":"2832_CR21","doi-asserted-by":"publisher","first-page":"1020","DOI":"10.1109\/TIP.2011.2179305","volume":"21","author":"K Chen","year":"2011","unstructured":"Chen, K., Lorenz, D.A.: Image sequence interpolation based on optical flow, segmentation, and optimal control. IEEE Trans. Image Process. 21(3), 1020\u20131030 (2011)","journal-title":"IEEE Trans. Image Process."},{"key":"2832_CR22","doi-asserted-by":"crossref","unstructured":"Werlberger, M., Pock, T., Unger, M., Bischof, H.: Optical flow guided TV-L1 video interpolation and restoration. Springer EMMCVPR pp. 273\u2013286 (2011)","DOI":"10.1007\/978-3-642-23094-3_20"},{"key":"2832_CR23","doi-asserted-by":"crossref","unstructured":"Jiang, H., Sun, D., Jampani, V., Yang, M.-H., Learned-Miller, E., Kautz, J.: Super SloMo: High quality estimation of multiple intermediate frames for video interpolation. IEEE CVPR pp. 9000\u20139008 (2018)","DOI":"10.1109\/CVPR.2018.00938"},{"key":"2832_CR24","doi-asserted-by":"crossref","unstructured":"Niklaus, S., Mai, L., Liu, F.: Video frame interpolation via adaptive convolution. IEEE CVPR pp. 2270\u20132279 (2017)","DOI":"10.1109\/CVPR.2017.244"},{"key":"2832_CR25","doi-asserted-by":"crossref","unstructured":"Liu, Z., Yeh, R.A., Tang, X., Liu, Y., Agarwala, A.: Video frame synthesis using deep voxel flow. IEEE ICCV pp. 4473\u20134481 (2017)","DOI":"10.1109\/ICCV.2017.478"},{"key":"2832_CR26","doi-asserted-by":"crossref","unstructured":"Liu, Y.-L., Liao, Y.-T., Lin, Y.-Y., Chuang, Y.-Y.: Deep video frame interpolation using cyclic frame generation. AAAI pp. 8794\u20138802 (2019)","DOI":"10.1609\/aaai.v33i01.33018794"},{"issue":"5","key":"2832_CR27","doi-asserted-by":"publisher","first-page":"619","DOI":"10.3390\/sym11050619","volume":"11","author":"H-E Ahn","year":"2019","unstructured":"Ahn, H.-E., Jeong, J., Kim, J.W.: A fast 4k video frame interpolation using a hybrid task-based convolutional neural network. MDPI Symmetry 11(5), 619 (2019)","journal-title":"MDPI Symmetry"},{"key":"2832_CR28","doi-asserted-by":"crossref","unstructured":"Xiang, X., Tian, Y., Zhang, Y., Fu, Y., Allebach, J.P., Xu, C.: Zooming Slow-Mo: Fast and accurate one-stage space-time video super-resolution. IEEE CVPR pp. 3367\u20133376 (2020)","DOI":"10.1109\/CVPR42600.2020.00343"},{"key":"2832_CR29","doi-asserted-by":"crossref","unstructured":"Bao, W., Lai, W.-S., Ma, C., Zhang, X., Gao, Z., Yang, M.-H.: Depth-aware video frame interpolation. IEEE CVPR pp. 3698\u20133707 (2019)","DOI":"10.1109\/CVPR.2019.00382"},{"issue":"3","key":"2832_CR30","doi-asserted-by":"publisher","first-page":"933","DOI":"10.1109\/TPAMI.2019.2941941","volume":"43","author":"W Bao","year":"2021","unstructured":"Bao, W., Lai, W.-S., Zhang, X., Gao, Z., Yang, M.-H.: MEMC-Net: Motion estimation and motion compensation driven neural network for video interpolation and enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 43(3), 933\u2013948 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2832_CR31","doi-asserted-by":"crossref","unstructured":"Wu, Y., Wen, Q., Chen, Q.: Optimizing video prediction via video frame interpolation. IEEE CVPR pp. 17814\u201317823 (2022)","DOI":"10.1109\/CVPR52688.2022.01729"},{"key":"2832_CR32","doi-asserted-by":"crossref","unstructured":"Shi, Z., Xu, X., Liu, X., Chen, J., Yang, M.-H.: Video frame interpolation transformer. IEEE CVPR pp. 17482\u201317491 (2022)","DOI":"10.1109\/CVPR52688.2022.01696"},{"key":"2832_CR33","doi-asserted-by":"crossref","unstructured":"Sim, H., Oh, J., Kim, M.: Xvfi: extreme video frame interpolation. ICCV pp. 14489\u201314498 (2021)","DOI":"10.1109\/ICCV48922.2021.01422"},{"key":"2832_CR34","unstructured":"Srivastava, N., Mansimov, E., Salakhudinov, R.: Unsupervised learning of video representations using LSTMs. ICML pp. 843\u2013852 (2015)"},{"key":"2832_CR35","unstructured":"Lotter, W., Kreiman, G., Cox, D.: Deep predictive coding networks for video prediction and unsupervised learning, ICLR (2017)"},{"key":"2832_CR36","unstructured":"Villegas, R., Yang, J., Hong, S., Lin, X., Lee, H.: Decomposing motion and content for natural video sequence prediction,\u2019 ICLR, (2017)"},{"key":"2832_CR37","unstructured":"Wang, Y., Long, M., Wang, J., Gao, Z., Yu, P.S.: PredRNN: Recurrent neural networks for predictive learning using spatiotemporal LSTMs. NIPS pp. 879\u2013888 (2017)"},{"key":"2832_CR38","unstructured":"Wang, Y., Gao, Z., Long, M., Wang, J., Yu, P.S.: PredRNN++: Towards a resolution of the deep-in-time dilemma in spatiotemporal predictive learning. ICML pp. 5123\u20135132 (2018)"},{"key":"2832_CR39","doi-asserted-by":"crossref","unstructured":"Liu, B., Chen, Y., Liu, S., Kim, H.-S.: Deep learning in latent space for video prediction and compression. IEEE CVPR pp. 701\u2013710 (2021)","DOI":"10.1109\/CVPR46437.2021.00076"},{"key":"2832_CR40","doi-asserted-by":"crossref","unstructured":"Wu, Y., Gao, R., Park, J., Chen, Q.: Future video synthesis with object motion prediction, IEEE CVPR, (2020)","DOI":"10.1109\/CVPR42600.2020.00558"},{"key":"2832_CR41","unstructured":"Kumar, M., Babaeizadeh, M., Erhan, D., Finn, C., Levine, S., Dinh, L., Kingma, D.: Videoflow: a conditional flow-based model for stochastic video generation, ICLR, (2020)"},{"issue":"5","key":"2832_CR42","doi-asserted-by":"publisher","first-page":"1053","DOI":"10.1109\/TPAMI.2019.2951667","volume":"42","author":"R Szeto","year":"2019","unstructured":"Szeto, R., Sun, X., Lu, K., Corso, J.J.: A temporally-aware interpolation network for video frame inpainting. IEEE Trans. Pattern Anal. Mach. Intell. 42(5), 1053\u20131068 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2832_CR43","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., van der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. IEEE CVPR pp. 2261\u20132269 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"issue":"3","key":"2832_CR44","doi-asserted-by":"publisher","first-page":"819","DOI":"10.1109\/TMM.2019.2932564","volume":"22","author":"S Agethen","year":"2019","unstructured":"Agethen, S., Hsu, W.H.: Deep multi-kernel convolutional LSTM networks and an attention-based mechanism for videos. IEEE Trans. Multimedia 22(3), 819\u2013829 (2019)","journal-title":"IEEE Trans. Multimedia"},{"key":"2832_CR45","doi-asserted-by":"crossref","unstructured":"Schuldt, C., Laptev, I., Caputo, B.: Recognizing human actions: a local SVM approach. IEEE ICPR pp. 32\u201336 (2004)","DOI":"10.1109\/ICPR.2004.1334462"},{"key":"2832_CR46","doi-asserted-by":"crossref","unstructured":"Kuehne, H., Jhuang, H., Garrote, E., Poggio, T., Serre, T.: HMDB: a large video database for human motion recognition. IEEE CCV pp. 2556\u20132563 (2011)","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"2832_CR47","unstructured":"Soomro, K., Zamir, A.\u00a0R., Shah, M.: UCF101: a dataset of 101 human actions classes from videos in the wild"},{"key":"2832_CR48","unstructured":"Bengio, S., Vinyals, O., Jaitly, N., Shazeer, N.: Scheduled sampling for sequence prediction with recurrent neural networks. NIPS pp. 1171\u20131179 (2015)"},{"key":"2832_CR49","unstructured":"Mathieu, M., Couprie, C., LeCun, Y.: Deep multi-scale video prediction beyond mean square error, ICLR (2016)"},{"key":"2832_CR50","unstructured":"Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization, ICLR (2015)"},{"issue":"4","key":"2832_CR51","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02832-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-02832-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02832-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,23]],"date-time":"2024-01-23T19:13:38Z","timestamp":1706037218000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-02832-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,11]]},"references-count":51,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,2]]}},"alternative-id":["2832"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-02832-y","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,4,11]]},"assertion":[{"value":"25 February 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 April 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declaration"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}