{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T18:34:49Z","timestamp":1772908489501,"version":"3.50.1"},"reference-count":37,"publisher":"Association for Computing Machinery (ACM)","issue":"1","license":[{"start":{"date-parts":[[2020,1,16]],"date-time":"2020-01-16T00:00:00Z","timestamp":1579132800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Korea government","award":["2017-0-01780"],"award-info":[{"award-number":["2017-0-01780"]}]},{"name":"Institute for Information 8 communications Technology Promotion"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Graph."],"published-print":{"date-parts":[[2020,2,29]]},"abstract":"<jats:p>Video stabilization is a fundamental and important technique for higher quality videos. Prior works have extensively explored video stabilization, but most of them involve cropping of the frame boundaries and introduce moderate levels of distortion. We present a novel deep approach to video stabilization that can generate video frames without cropping and low distortion. The proposed framework utilizes frame interpolation techniques to generate in between frames, leading to reduced inter-frame jitter. Once applied in an iterative fashion, the stabilization effect becomes stronger. A major advantage is that our framework is end-to-end trainable in an unsupervised manner. In addition, our method is able to run in near real-time (15 fps). To the best of our knowledge, this is the first work to propose an unsupervised deep approach to full-frame video stabilization. We show the advantages of our method through quantitative and qualitative evaluations comparing to the state-of-the-art methods.<\/jats:p>","DOI":"10.1145\/3363550","type":"journal-article","created":{"date-parts":[[2020,1,17]],"date-time":"2020-01-17T04:20:37Z","timestamp":1579234837000},"page":"1-9","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":78,"title":["Deep Iterative Frame Interpolation for Full-frame Video Stabilization"],"prefix":"10.1145","volume":"39","author":[{"given":"Jinsoo","family":"Choi","sequence":"first","affiliation":[{"name":"KAIST, Daejeon, Republic of Korea"}]},{"given":"In So","family":"Kweon","sequence":"additional","affiliation":[{"name":"KAIST, Daejeon, Republic of Korea"}]}],"member":"320","published-online":{"date-parts":[[2020,1,16]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.12413"},{"key":"e_1_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10593-2_20"},{"key":"e_1_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2001.991019"},{"key":"e_1_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1467-8659.2008.01326.x"},{"key":"e_1_2_1_5_1","volume-title":"Proceedings of the ACM International Conference on Multimedia. 27--36","author":"Michael","unstructured":"Michael L. Gleicher and Feng Liu. 2007. Re-cinematography: Improving the camera dynamics of casual video . In Proceedings of the ACM International Conference on Multimedia. 27--36 . Michael L. Gleicher and Feng Liu. 2007. Re-cinematography: Improving the camera dynamics of casual video. In Proceedings of the ACM International Conference on Multimedia. 27--36."},{"key":"e_1_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/2231816.2231824"},{"key":"e_1_2_1_7_1","volume-title":"Proceedings of the Conference on Advances in Neural Information Processing Systems (NIPS\u201915)","author":"Goroshin Ross","year":"2015","unstructured":"Ross Goroshin , Michael F. Mathieu , and Yann LeCun . 2015 . Learning to linearize under uncertainty . In Proceedings of the Conference on Advances in Neural Information Processing Systems (NIPS\u201915) . 1234--1242. Ross Goroshin, Michael F. Mathieu, and Yann LeCun. 2015. Learning to linearize under uncertainty. In Proceedings of the Conference on Advances in Neural Information Processing Systems (NIPS\u201915). 1234--1242."},{"key":"e_1_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCPhot.2012.6215213"},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2011.5995525"},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_2_1_11_1","article-title":"Encoding shaky videos by integrating efficient video stabilization","volume":"29","author":"Huang Hua","year":"2018","unstructured":"Hua Huang , Xiao-Xiang Wei , and Lei Zhang . 2018 . Encoding shaky videos by integrating efficient video stabilization . IEEE Trans. Circ. Syst. Vid. Technol. 29 , 5 (2018). Hua Huang, Xiao-Xiang Wei, and Lei Zhang. 2018. Encoding shaky videos by integrating efficient video stabilization. IEEE Trans. Circ. Syst. Vid. Technol. 29, 5 (2018).","journal-title":"IEEE Trans. Circ. Syst. Vid. Technol."},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/1531326.1531350"},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/1899404.1899408"},{"key":"e_1_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.16"},{"key":"e_1_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2697759"},{"key":"e_1_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46466-4_48"},{"key":"e_1_2_1_18_1","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201912)","author":"Liu Shuaicheng","year":"2012","unstructured":"Shuaicheng Liu , Yinting Wang , Lu Yuan , Jiajun Bu , Ping Tan , and Jian Sun . 2012 . Video stabilization with a depth camera . In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201912) . 89--95. Shuaicheng Liu, Yinting Wang, Lu Yuan, Jiajun Bu, Ping Tan, and Jian Sun. 2012. Video stabilization with a depth camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201912). 89--95."},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/2461912.2461995"},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.536"},{"key":"e_1_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46466-4_26"},{"key":"e_1_2_1_22_1","volume-title":"Proceedings of the International Conference on Learning Representations (ICLR\u201916)","author":"Mathieu Michael","year":"2016","unstructured":"Michael Mathieu , Camille Couprie , and Yann LeCun . 2016 . Deep multi-scale video prediction beyond mean square error . In Proceedings of the International Conference on Learning Representations (ICLR\u201916) . Michael Mathieu, Camille Couprie, and Yann LeCun. 2016. Deep multi-scale video prediction beyond mean square error. In Proceedings of the International Conference on Learning Representations (ICLR\u201916)."},{"key":"e_1_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2006.141"},{"key":"e_1_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298747"},{"key":"e_1_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00183"},{"key":"e_1_2_1_26_1","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"1","author":"Niklaus Simon","year":"2017","unstructured":"Simon Niklaus , Long Mai , and Feng Liu . 2017 a. Video frame interpolation via adaptive convolution . In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) , Vol. 1 . 3. Simon Niklaus, Long Mai, and Feng Liu. 2017a. Video frame interpolation via adaptive convolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Vol. 1. 3."},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.37"},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2015.7139474"},{"key":"e_1_2_1_29_1","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201916)","author":"Perazzi F.","unstructured":"F. Perazzi , J. Pont-Tuset , B. McWilliams , L. Van Gool , M. Gross , and A. Sorkine-Hornung . 2016. A benchmark dataset and evaluation methodology for video object segmentation . In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201916) . F. Perazzi, J. Pont-Tuset, B. McWilliams, L. Van Gool, M. Gross, and A. Sorkine-Hornung. 2016. A benchmark dataset and evaluation methodology for video object segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR\u201916)."},{"key":"e_1_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"e_1_2_1_31_1","volume-title":"Very deep convolutional networks for large-scale image recognition. Retrieved from arXiv preprint arXiv:1409.1556","author":"Simonyan Karen","year":"2014","unstructured":"Karen Simonyan and Andrew Zisserman . 2014. Very deep convolutional networks for large-scale image recognition. Retrieved from arXiv preprint arXiv:1409.1556 ( 2014 ). Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networks for large-scale image recognition. Retrieved from arXiv preprint arXiv:1409.1556 (2014)."},{"key":"e_1_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2009.5459270"},{"key":"e_1_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00931"},{"key":"e_1_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2018.2884280"},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2013.11"},{"key":"e_1_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.13566"},{"key":"e_1_2_1_37_1","volume-title":"Free-form image inpainting with gated convolution. Retrieved from arXiv preprint arXiv:1806.03589","author":"Yu Jiahui","year":"2018","unstructured":"Jiahui Yu , Zhe Lin , Jimei Yang , Xiaohui Shen , Xin Lu , and Thomas S Huang . 2018. Free-form image inpainting with gated convolution. Retrieved from arXiv preprint arXiv:1806.03589 ( 2018 ). Jiahui Yu, Zhe Lin, Jimei Yang, Xiaohui Shen, Xin Lu, and Thomas S Huang. 2018. Free-form image inpainting with gated convolution. Retrieved from arXiv preprint arXiv:1806.03589 (2018)."},{"key":"e_1_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2013.298"}],"container-title":["ACM Transactions on Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3363550","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3363550","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T23:44:25Z","timestamp":1750203865000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3363550"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,1,16]]},"references-count":37,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2020,2,29]]}},"alternative-id":["10.1145\/3363550"],"URL":"https:\/\/doi.org\/10.1145\/3363550","relation":{},"ISSN":["0730-0301","1557-7368"],"issn-type":[{"value":"0730-0301","type":"print"},{"value":"1557-7368","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,1,16]]},"assertion":[{"value":"2019-04-01","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2019-09-01","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2020-01-16","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}