{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,27]],"date-time":"2026-02-27T11:19:15Z","timestamp":1772191155893,"version":"3.50.1"},"reference-count":32,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2022,9,18]],"date-time":"2022-09-18T00:00:00Z","timestamp":1663459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,9,18]],"date-time":"2022-09-18T00:00:00Z","timestamp":1663459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Pattern Anal Applic"],"published-print":{"date-parts":[[2023,2]]},"DOI":"10.1007\/s10044-022-01106-y","type":"journal-article","created":{"date-parts":[[2022,9,18]],"date-time":"2022-09-18T19:02:22Z","timestamp":1663527742000},"page":"343-355","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Correlation-based and content-enhanced network for video style transfer"],"prefix":"10.1007","volume":"26","author":[{"given":"Honglin","family":"Lin","sequence":"first","affiliation":[]},{"given":"Mengmeng","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jiaxin","family":"Kou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,18]]},"reference":[{"key":"1106_CR1","doi-asserted-by":"crossref","unstructured":"Gatys LA, Ecker AS, Bethge M (2016) Image style transfer using convolutional neural networks. In: IEEE conference on Computer Vision and Pattern Recognition, pp. 2414\u20132423","DOI":"10.1109\/CVPR.2016.265"},{"key":"1106_CR2","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556"},{"key":"1106_CR3","doi-asserted-by":"crossref","unstructured":"Ruder M, Dosovitskiy A, Brox T (2016) Artistic style transfer for videos. In: German Conference on Pattern Recognition, pp. 26\u201336. Springer","DOI":"10.1007\/978-3-319-45886-1_3"},{"key":"1106_CR4","doi-asserted-by":"crossref","unstructured":"Huang H, Wang H, Luo W, Ma L, Jiang W, Zhu X, Li Z, Liu W (2017) Real-time neural style transfer for videos. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 783\u2013791","DOI":"10.1109\/CVPR.2017.745"},{"key":"1106_CR5","doi-asserted-by":"crossref","unstructured":"Gupta A, Johnson J, Alahi A, Fei-Fei L (2017) Characterizing and improving stability in neural style transfer. In: IEEE International Conference on Computer Vision, pp. 4067\u20134076","DOI":"10.1109\/ICCV.2017.438"},{"key":"1106_CR6","doi-asserted-by":"crossref","unstructured":"Chen X, Zhang Y, Wang Y, Shu H, Xu C, Xu C (2020) Optical flow distillation: towards efficient and stable video style transfer. In: European Conference on Computer Vision, pp. 614\u2013630. Springer","DOI":"10.1007\/978-3-030-58539-6_37"},{"key":"1106_CR7","doi-asserted-by":"crossref","unstructured":"Chen D, Liao J, Yuan L, Yu N, Hua G (2017) Coherent online video style transfer. In: IEEE International Conference on Computer Vision, pp. 1105\u20131114","DOI":"10.1109\/ICCV.2017.126"},{"key":"1106_CR8","doi-asserted-by":"publisher","first-page":"9125","DOI":"10.1109\/TIP.2020.3024018","volume":"29","author":"W Wang","year":"2020","unstructured":"Wang W, Yang S, Xu J, Liu J (2020) Consistent video style transfer via relaxation and regularization. IEEE Trans Image Process 29:9125\u20139139","journal-title":"IEEE Trans Image Process"},{"key":"1106_CR9","doi-asserted-by":"crossref","unstructured":"Deng Y, Tang F, Dong W, Huang H, Ma C, Xu C (2020) Arbitrary video style transfer via multi-channel correlation. arXiv preprint arXiv:2009.08003","DOI":"10.1145\/3394171.3414015"},{"key":"1106_CR10","doi-asserted-by":"crossref","unstructured":"Johnson J, Alahi A, Fei-Fei L (2016) Perceptual losses for real-time style transfer and super-resolution. In: European Conference on Computer Vision, pp. 694\u2013711. Springer","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"1106_CR11","doi-asserted-by":"crossref","unstructured":"Chen D, Yuan L, Liao J, Yu N, Hua G (2017) Stylebank: an explicit representation for neural image style transfer. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1897\u20131906","DOI":"10.1109\/CVPR.2017.296"},{"key":"1106_CR12","unstructured":"Dumoulin V, Shlens J, Kudlur M (2016) A learned representation for artistic style. arXiv preprint arXiv:1610.07629"},{"key":"1106_CR13","doi-asserted-by":"crossref","unstructured":"Huang X, Belongie S (2017) Arbitrary style transfer in real-time with adaptive instance normalization. In: IEEE International Conference on Computer Vision, pp. 1501\u20131510","DOI":"10.1109\/ICCV.2017.167"},{"key":"1106_CR14","unstructured":"Li Y, Fang C, Yang J, Wang Z, Lu X, Yang M-H (2017) Universal style transfer via feature transforms. arXiv preprint arXiv:1705.08086"},{"key":"1106_CR15","doi-asserted-by":"crossref","unstructured":"Li X, Liu S, Kautz J, Yang M-H (2019) Learning linear transformations for fast image and video style transfer. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 3809\u20133817","DOI":"10.1109\/CVPR.2019.00393"},{"key":"1106_CR16","doi-asserted-by":"crossref","unstructured":"Park DY, Lee KH (2019) Arbitrary style transfer with style-attentional networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 5880\u20135888","DOI":"10.1109\/CVPR.2019.00603"},{"key":"1106_CR17","doi-asserted-by":"crossref","unstructured":"Yao Y, Ren J, Xie X, Liu W, Liu Y-J, Wang J (2019) Attention-aware multi-stroke style transfer. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1467\u20131475","DOI":"10.1109\/CVPR.2019.00156"},{"key":"1106_CR18","doi-asserted-by":"crossref","unstructured":"Liu S, Lin T, He D, Li F, Wang M, Li X, Sun Z, Li Q, Ding E (2021) Adaattn: revisit attention mechanism in arbitrary neural style transfer. In: IEEE International Conference on Computer Vision, pp. 6649\u20136658","DOI":"10.1109\/ICCV48922.2021.00658"},{"key":"1106_CR19","doi-asserted-by":"crossref","unstructured":"Lai W-S, Huang J-B, Wang O, Shechtman E, Yumer E, Yang M-H (2018) Learning blind video temporal consistency. In: European Conference on Computer Vision, pp. 170\u2013185","DOI":"10.1007\/978-3-030-01267-0_11"},{"key":"1106_CR20","doi-asserted-by":"crossref","unstructured":"Gao W, Li Y, Yin Y, Yang M-H (2020) Fast video multi-style transfer. In: IEEE Winter Conference on Applications of Computer Vision, pp. 3222\u20133230","DOI":"10.1109\/WACV45572.2020.9093420"},{"key":"1106_CR21","unstructured":"Wang T-C, Liu M-Y, Zhu J-Y, Liu G, Tao A, Kautz J, Catanzaro, B. (2018) Video-to-video synthesis. arXiv preprint arXiv:1808.06601"},{"key":"1106_CR22","doi-asserted-by":"crossref","unstructured":"Zhou Y, Xu X, Shen F, Gao L, Lu H, Shen, HT (2020) Temporal denoising mask synthesis network for learning blind video temporal consistency. In: ACM International Conference on Multimedia, pp. 475\u2013483","DOI":"10.1145\/3394171.3413788"},{"key":"1106_CR23","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"1106_CR24","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser \u0141, Polosukhin I (2017) Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008"},{"key":"1106_CR25","unstructured":"Arjovsky M, Chintala S, Bottou L (2017) Wasserstein generative adversarial networks. In: International Conference on Machine Learning, pp. 214\u2013223. PMLR"},{"key":"1106_CR26","doi-asserted-by":"crossref","unstructured":"Li Y, Wang N, Liu J, Hou X (2017) Demystifying neural style transfer. arXiv preprint arXiv:1701.01036","DOI":"10.24963\/ijcai.2017\/310"},{"key":"1106_CR27","doi-asserted-by":"crossref","unstructured":"Kolkin N, Salavon J, Shakhnarovich G. (2019) Style transfer by relaxed optimal transport and self-similarity. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 10051\u201310060","DOI":"10.1109\/CVPR.2019.01029"},{"key":"1106_CR28","doi-asserted-by":"publisher","first-page":"2501","DOI":"10.1109\/TIP.2021.3052709","volume":"30","author":"K Xu","year":"2021","unstructured":"Xu K, Wen L, Li G, Qi H, Bo L, Huang Q (2021) Learning self-supervised space-time cnn for fast video style transfer. IEEE Trans Image Process 30:2501\u20132512","journal-title":"IEEE Trans Image Process"},{"key":"1106_CR29","doi-asserted-by":"crossref","unstructured":"Lin T-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, Doll\u00e1r P, Zitnick CL (2014) Microsoft coco: common objects in context. In: European Conference on Computer Vision, pp. 740\u2013755. Springer","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"1106_CR30","unstructured":"Saleh B, Elgammal A (2015) Large-scale classification of fine-art paintings: learning the right metric on the right feature. arXiv preprint arXiv:1505.00855"},{"key":"1106_CR31","doi-asserted-by":"crossref","unstructured":"Butler DJ, Wulff J, Stanley GB, Black MJ (2012) A naturalistic open source movie for optical flow evaluation. In: European Conference on Computer Vision, pp. 611\u2013625. Springer","DOI":"10.1007\/978-3-642-33783-3_44"},{"key":"1106_CR32","unstructured":"Pont-Tuset J, Perazzi F, Caelles S, Arbel\u00e1ez P, Sorkine-Hornung A, Gool LV (2018) The 2017 davis challenge on video object segmentation. arXiv preprint arXiv:1704.00675"}],"container-title":["Pattern Analysis and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-022-01106-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10044-022-01106-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-022-01106-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,4]],"date-time":"2023-02-04T03:25:37Z","timestamp":1675481137000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10044-022-01106-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,9,18]]},"references-count":32,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2023,2]]}},"alternative-id":["1106"],"URL":"https:\/\/doi.org\/10.1007\/s10044-022-01106-y","relation":{},"ISSN":["1433-7541","1433-755X"],"issn-type":[{"value":"1433-7541","type":"print"},{"value":"1433-755X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,9,18]]},"assertion":[{"value":"9 February 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 August 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 September 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}