{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,18]],"date-time":"2026-01-18T14:44:54Z","timestamp":1768747494536,"version":"3.49.0"},"reference-count":35,"publisher":"Association for Computing Machinery (ACM)","issue":"4","license":[{"start":{"date-parts":[[2020,8,12]],"date-time":"2020-08-12T00:00:00Z","timestamp":1597190400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100011103","name":"European Commission","doi-asserted-by":"publisher","award":["CZ.02.1.01\/0.0\/0.0\/16_019\/0000765"],"award-info":[{"award-number":["CZ.02.1.01\/0.0\/0.0\/16_019\/0000765"]}],"id":[{"id":"10.13039\/100011103","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Grant Agency of the Czech Technical University in Prague","award":["SGS19\/179\/OHK3\/3T\/13"],"award-info":[{"award-number":["SGS19\/179\/OHK3\/3T\/13"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Graph."],"published-print":{"date-parts":[[2020,8,31]]},"abstract":"<jats:p>In this paper, we present a learning-based method to the keyframe-based video stylization that allows an artist to propagate the style from a few selected keyframes to the rest of the sequence. Its key advantage is that the resulting stylization is semantically meaningful, i.e., specific parts of moving objects are stylized according to the artist's intention. In contrast to previous style transfer techniques, our approach does not require any lengthy pre-training process nor a large training dataset. We demonstrate how to train an appearance translation network from scratch using only a few stylized exemplars while implicitly preserving temporal consistency. This leads to a video stylization framework that supports real-time inference, parallel processing, and random access to an arbitrary output frame. It can also merge the content from multiple keyframes without the need to perform an explicit blending operation. We demonstrate its practical utility in various interactive scenarios, where the user paints over a selected keyframe and sees her style transferred to an existing recorded sequence or a live video stream.<\/jats:p>","DOI":"10.1145\/3386569.3392453","type":"journal-article","created":{"date-parts":[[2020,8,12]],"date-time":"2020-08-12T11:44:27Z","timestamp":1597232667000},"update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":60,"title":["Interactive video stylization using few-shot patch-based training"],"prefix":"10.1145","volume":"39","author":[{"given":"Ond\u0159ej","family":"Texler","sequence":"first","affiliation":[{"name":"Czech Technical University in Prague"}]},{"given":"David","family":"Futschik","sequence":"additional","affiliation":[{"name":"Czech Technical University in Prague"}]},{"given":"Michal","family":"ku\u010dera","sequence":"additional","affiliation":[{"name":"Czech Technical University in Prague"}]},{"given":"Ond\u0159ej","family":"jamri\u0161ka","sequence":"additional","affiliation":[{"name":"Czech Technical University in Prague"}]},{"given":"\u0160\u00e1rka","family":"Sochorov\u00e1","sequence":"additional","affiliation":[{"name":"Czech Technical University in Prague"}]},{"given":"Menclei","family":"Chai","sequence":"additional","affiliation":[{"name":"Snap Inc."}]},{"given":"Sergey","family":"Tulyakov","sequence":"additional","affiliation":[{"name":"Snap Inc."}]},{"given":"Daniel","family":"S\u00ddkora","sequence":"additional","affiliation":[{"name":"Czech Technical University in Prague"}]}],"member":"320","published-online":{"date-parts":[[2020,8,12]]},"reference":[{"key":"e_1_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/2461912.2461929"},{"key":"e_1_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/1073204.1073272"},{"key":"e_1_2_2_3_1","volume-title":"Proceedings of IEEE International Conference on Computer Vision. 5933--5942","author":"Chan Caroline","unstructured":"Caroline Chan, Shiry Ginosar, Tinghui Zhou, and Alexei A. Efros. 2019. Everybody Dance Now. In Proceedings of IEEE International Conference on Computer Vision. 5933--5942."},{"key":"e_1_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.126"},{"key":"e_1_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2013.316"},{"key":"e_1_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/2897824.2925948"},{"key":"e_1_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3072959.3073660"},{"key":"e_1_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.12407"},{"key":"e_1_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-018-1474-1"},{"key":"e_1_2_2_10_1","volume-title":"Proceedings of the ACM\/EG Expressive Symposium. 33--42","author":"Futschik David","year":"2019","unstructured":"David Futschik, Menglei Chai, Chen Cao, Chongyang Ma, Aleksei Stoliar, Sergey Korolev, Sergey Tulyakov, Michal Ku\u010dera, and Daniel S\u00fdkora. 2019. Real-Time Patch-Based Stylization of Portraits Using Generative Adversarial Network. In Proceedings of the ACM\/EG Expressive Symposium. 33--42."},{"key":"e_1_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.265"},{"key":"e_1_2_2_12_1","volume-title":"Image Analogies. In SIGGRAPH Conference Proceedings. 327--340","author":"Hertzmann Aaron","unstructured":"Aaron Hertzmann, Charles E. Jacobs, Nuria Oliver, Brian Curless, and David H. Salesin. 2001. Image Analogies. In SIGGRAPH Conference Proceedings. 327--340."},{"key":"e_1_2_2_13_1","volume-title":"Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 5967--5976","author":"Isola Phillip","unstructured":"Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A. Efros. 2017. Image-to-Image Translation with Conditional Adversarial Networks. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 5967--5976."},{"key":"e_1_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306346.3323006"},{"key":"e_1_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00452"},{"key":"e_1_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01267-0_11"},{"key":"e_1_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46487-9_43"},{"key":"e_1_2_2_18_1","volume-title":"Xiaolong Wang, Jan Kautz, and Ming-Hsuan Yang.","author":"Li Xueting","year":"2019","unstructured":"Xueting Li, Sifei Liu, Shalini De Mello, Xiaolong Wang, Jan Kautz, and Ming-Hsuan Yang. 2019. Joint-task Self-supervised Learning for Temporal Correspondence. In Advances in Neural Information Processing Systems. 317--327."},{"key":"e_1_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.01065"},{"key":"e_1_2_2_20_1","volume-title":"Eric Frank, Alex Sergeev, and Jason Yosinski.","author":"Liu Rosanne","year":"2018","unstructured":"Rosanne Liu, Joel Lehman, Piero Molino, Felipe Petroski Such, Eric Frank, Alex Sergeev, and Jason Yosinski. 2018. An intriguing failing of convolutional neural networks and the CoordConv solution. In Advances in Neural Information Processing Systems. 9628--9639."},{"key":"e_1_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00367"},{"key":"e_1_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-018-1089-z"},{"key":"e_1_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01237-3_43"},{"key":"e_1_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00467"},{"key":"e_1_2_2_25_1","volume-title":"Regenerative Morphing. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. 615--622","author":"Shechtman Eli","unstructured":"Eli Shechtman, Alex Rav-Acha, Michal Irani, and Steven M. Seitz. 2010. Regenerative Morphing. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. 615--622."},{"key":"e_1_2_2_26_1","volume-title":"Proceedings of IEEE International Conference on Computer Vision. 4492--4501","author":"Shocher Assaf","year":"2019","unstructured":"Assaf Shocher, Shai Bagon, Phillip Isola, and Michal Irani. 2019. InGAN: Capturing and Remapping the \"DNA\" of a Natural Image. In Proceedings of IEEE International Conference on Computer Vision. 4492--4501."},{"key":"e_1_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00329"},{"key":"e_1_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/1572614.1572619"},{"key":"e_1_2_2_29_1","volume-title":"Proceedings of International Conference on International Conference on Machine Learning. 1349--1357","author":"Ulyanov Dmitry","year":"2016","unstructured":"Dmitry Ulyanov, Vadim Lebedev, Andrea Vedaldi, and Victor Lempitsky. 2016. Texture Networks: Feed-forward Synthesis of Textures and Stylized Images. In Proceedings of International Conference on International Conference on Machine Learning. 1349--1357."},{"key":"e_1_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00159"},{"key":"e_1_2_2_31_1","unstructured":"Ting-Chun Wang Ming-Yu Liu Andrew Tao Guilin Liu Jan Kautz and Bryan Catanzaro. 2019b. Few-shot Video-to-Video Synthesis. In Advances in Neural Information Processing Systems. 5014--5025."},{"key":"e_1_2_2_32_1","unstructured":"Ting-Chun Wang Ming-Yu Liu Jun-Yan Zhu Guilin Liu Andrew Tao Jan Kautz and Bryan Catanzaro. 2018a. Video-to-Video Synthesis. In Advances in Neural Information Processing Systems. 1144--1156."},{"key":"e_1_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00917"},{"key":"e_1_2_2_34_1","volume-title":"Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2561--2571","author":"Wang Xiaolong","unstructured":"Xiaolong Wang, Allan Jabri, and Alexei A. Efros. 2019a. Learning Correspondence From the Cycle-Consistency of Time. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2561--2571."},{"key":"e_1_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3197517.3201285"}],"container-title":["ACM Transactions on Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3386569.3392453","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3386569.3392453","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,25]],"date-time":"2025-06-25T05:40:41Z","timestamp":1750830041000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3386569.3392453"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,8,12]]},"references-count":35,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2020,8,31]]}},"alternative-id":["10.1145\/3386569.3392453"],"URL":"https:\/\/doi.org\/10.1145\/3386569.3392453","relation":{},"ISSN":["0730-0301","1557-7368"],"issn-type":[{"value":"0730-0301","type":"print"},{"value":"1557-7368","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,8,12]]},"assertion":[{"value":"2020-08-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}