{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,4]],"date-time":"2025-11-04T11:12:37Z","timestamp":1762254757502,"version":"3.41.0"},"reference-count":64,"publisher":"Association for Computing Machinery (ACM)","issue":"5","license":[{"start":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T00:00:00Z","timestamp":1705881600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Multimedia Comput. Commun. Appl."],"published-print":{"date-parts":[[2024,5,31]]},"abstract":"<jats:p>The objective of arbitrary style transfer is to apply a given artistic or photo-realistic style to a target image. Although current methods have shown some success in transferring style, arbitrary style transfer still has several issues, including content leakage. Embedding an artistic style can result in unintended changes to the image content. This article proposes an iterative framework called Restorable Arbitrary Style Transfer (RAST) to effectively ensure content preservation and mitigate potential alterations to the content information. RAST can transmit both content and style information through multi-restorations and balance the content-style tradeoff in stylized images using the image restoration accuracy. To ensure RAST\u2019s effectiveness, we introduce two novel loss functions: multi-restoration loss and style difference loss. We also propose a new quantitative evaluation method to assess content preservation and style embedding performance. Experimental results show that RAST outperforms state-of-the-art methods in generating stylized images that preserve content and embed style accurately.<\/jats:p>","DOI":"10.1145\/3638770","type":"journal-article","created":{"date-parts":[[2023,12,30]],"date-time":"2023-12-30T15:57:33Z","timestamp":1703951853000},"page":"1-21","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":8,"title":["RAST: Restorable Arbitrary Style Transfer"],"prefix":"10.1145","volume":"20","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1885-6424","authenticated-orcid":false,"given":"Yingnan","family":"Ma","sequence":"first","affiliation":[{"name":"University of Alberta, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4574-4815","authenticated-orcid":false,"given":"Chenqiu","family":"Zhao","sequence":"additional","affiliation":[{"name":"University of Alberta, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-5793-8100","authenticated-orcid":false,"given":"Bingran","family":"Huang","sequence":"additional","affiliation":[{"name":"University of Alberta, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-8828-9488","authenticated-orcid":false,"given":"Xudong","family":"Li","sequence":"additional","affiliation":[{"name":"University of Alberta, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7695-4148","authenticated-orcid":false,"given":"Anup","family":"Basu","sequence":"additional","affiliation":[{"name":"University of Alberta, Canada"}]}],"member":"320","published-online":{"date-parts":[[2024,1,22]]},"reference":[{"key":"e_1_3_1_2_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00092"},{"key":"e_1_3_1_3_2","doi-asserted-by":"publisher","DOI":"10.1145\/3475799"},{"key":"e_1_3_1_4_2","unstructured":"Haibo Chen Lei Zhao Zhizhong Wang Huiming Zhang Zhiwen Zuo Ailin Li Wei Xing and Dongming Lu. 2021. Artistic style transfer with internal-external learning and contrastive learning. Advances in Neural Information Processing Systems 34 (2021) 26561\u201326573."},{"key":"e_1_3_1_5_2","unstructured":"Tian Qi Chen and Mark Schmidt. 2016. Fast patch-based style transfer of arbitrary style. arXiv:1612.04337. Retrieved from https:\/\/arxiv.org\/abs\/1612.04337"},{"key":"e_1_3_1_6_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00916"},{"key":"e_1_3_1_7_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i2.16208"},{"key":"e_1_3_1_8_2","first-page":"11326","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Deng Yingying","year":"2022","unstructured":"Yingying Deng, Fan Tang, Weiming Dong, Chongyang Ma, Xingjia Pan, Lei Wang, and Changsheng Xu. 2022. StyTr2: Image style transfer with transformers. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 11326\u201311336."},{"key":"e_1_3_1_9_2","doi-asserted-by":"publisher","DOI":"10.1145\/3394171.3414015"},{"key":"e_1_3_1_10_2","unstructured":"Laurent Dinh David Krueger and Yoshua Bengio. 2015. NICE: Non-linear independent components estimation. arXiv:1410.8516. Retrieved from https:\/\/arxiv.org\/abs\/1410.8516"},{"key":"e_1_3_1_11_2","unstructured":"Vincent Dumoulin Jonathon Shlens and Manjunath Kudlur. 2016. A learned representation for artistic style. International Conference on Learning Representations."},{"key":"e_1_3_1_12_2","doi-asserted-by":"publisher","DOI":"10.1109\/tip.2017.2678168"},{"key":"e_1_3_1_13_2","doi-asserted-by":"publisher","DOI":"10.1145\/2897824.2925948"},{"key":"e_1_3_1_14_2","first-page":"553","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","author":"Frigo Oriel","year":"2016","unstructured":"Oriel Frigo, Neus Sabater, Julie Delon, and Pierre Hellier. 2016. Split and match: Example-based adaptive patch sampling for unsupervised style transfer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 553\u2013561."},{"key":"e_1_3_1_15_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.265"},{"key":"e_1_3_1_16_2","doi-asserted-by":"publisher","DOI":"10.1145\/966131.966133"},{"key":"e_1_3_1_17_2","article-title":"Generative adversarial nets","author":"Goodfellow Ian","year":"2014","unstructured":"Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems.","journal-title":"Proceedings of the 27th International Conference on Neural Information Processing Systems"},{"key":"e_1_3_1_18_2","article-title":"PAINT: Photo-realistic fashion design synthesis","author":"Gu Xiaoling","year":"2023","unstructured":"Xiaoling Gu, Jie Huang, Yongkang Wong, Jun Yu, Jianping Fan, Pai Peng, and Mohan S. Kankanhalli. 2023. PAINT: Photo-realistic fashion design synthesis. ACM Transactions on Multimedia Computing, Communications, and Applications 20, 2 (2023), 1\u201323.","journal-title":"ACM Transactions on Multimedia Computing, Communications, and Applications"},{"key":"e_1_3_1_19_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"e_1_3_1_20_2","doi-asserted-by":"publisher","DOI":"10.1145\/280814.280951"},{"key":"e_1_3_1_21_2","doi-asserted-by":"publisher","DOI":"10.1145\/383259.383295"},{"key":"e_1_3_1_22_2","article-title":"GANs trained by a two time-scale update rule converge to a local nash equilibrium","author":"Heusel Martin","year":"2017","unstructured":"Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. 2017. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the 31st International Conference on Neural Information Processing Systems.","journal-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems"},{"key":"e_1_3_1_23_2","first-page":"2722","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Ho Jonathan","year":"2019","unstructured":"Jonathan Ho, Xi Chen, Aravind Srinivas, Yan Duan, and Pieter Abbeel. 2019. Flow++: Improving flow-based generative models with variational dequantization and architecture design. In Proceedings of the International Conference on Machine Learning. PMLR, 2722\u20132730."},{"key":"e_1_3_1_24_2","first-page":"14609","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","author":"Hong Kibeom","year":"2021","unstructured":"Kibeom Hong, Seogkyu Jeon, Huan Yang, Jianlong Fu, and Hyeran Byun. 2021. Domain-aware universal style transfer. In Proceedings of the IEEE\/CVF International Conference on Computer Vision. 14609\u201314617."},{"key":"e_1_3_1_25_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.167"},{"key":"e_1_3_1_26_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"e_1_3_1_27_2","first-page":"21357","article-title":"ContraGAN: Contrastive learning for conditional image generation","author":"Kang Minguk","year":"2020","unstructured":"Minguk Kang and Jaesik Park. 2020. ContraGAN: Contrastive learning for conditional image generation. In Proceedings of the 34th International Conference on Neural Information Processing Systems. 21357\u201321369.","journal-title":"Proceedings of the 34th International Conference on Neural Information Processing Systems"},{"key":"e_1_3_1_28_2","doi-asserted-by":"publisher","unstructured":"Sergey Karayev Matthew Trentacoste Helen Han Aseem Agarwala Trevor Darrell Aaron Hertzmann and Holger Winnemoeller. 2014. Recognizing image style. In Proceedings of the British Machine Vision Conference 2014. 10.5244\/C.28.122","DOI":"10.5244\/C.28.122"},{"key":"e_1_3_1_29_2","first-page":"1857","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Kim Taeksoo","year":"2017","unstructured":"Taeksoo Kim, Moonsu Cha, Hyunsoo Kim, Jung Kwon Lee, and Jiwon Kim. 2017. Learning to discover cross-domain relations with generative adversarial networks. In Proceedings of the International Conference on Machine Learning. PMLR, 1857\u20131865."},{"key":"e_1_3_1_30_2","unstructured":"Diederik P. Kingma and Jimmy Ba. 2017. Adam: A method for stochastic optimization. arXiv:1412.6980. Retrieved from https:\/\/arxiv.org\/abs\/1412.6980"},{"key":"e_1_3_1_31_2","article-title":"Glow: Generative flow with invertible 1x1 convolutions","author":"Kingma Durk P.","year":"2018","unstructured":"Durk P. Kingma and Prafulla Dhariwal. 2018. Glow: Generative flow with invertible 1x1 convolutions. In Proceedings of the 32nd International Conference on Neural Information Processing Systems.","journal-title":"Proceedings of the 32nd International Conference on Neural Information Processing Systems"},{"key":"e_1_3_1_32_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00452"},{"key":"e_1_3_1_33_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.272"},{"key":"e_1_3_1_34_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00393"},{"key":"e_1_3_1_35_2","article-title":"Universal style transfer via feature transforms","author":"Li Yijun","year":"2017","unstructured":"Yijun Li, Chen Fang, Jimei Yang, Zhaowen Wang, Xin Lu, and Ming-Hsuan Yang. 2017. Universal style transfer via feature transforms. In Proceedings of the 31st International Conference on Neural Information Processing Systems.","journal-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems"},{"issue":"3","key":"e_1_3_1_36_2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3450525","article-title":"Distribution aligned multimodal and multi-domain image stylization","volume":"17","author":"Lin Minxuan","year":"2021","unstructured":"Minxuan Lin, Fan Tang, Weiming Dong, Xiao Li, Changsheng Xu, and Chongyang Ma. 2021. Distribution aligned multimodal and multi-domain image stylization. ACM Transactions on Multimedia Computing, Communications, and Applications 17, 3 (2021), 1\u201317.","journal-title":"ACM Transactions on Multimedia Computing, Communications, and Applications"},{"key":"e_1_3_1_37_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"e_1_3_1_38_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00658"},{"key":"e_1_3_1_39_2","unstructured":"Zhi-Song Liu Li-Wen Wang Wan-Chi Siu and Vicky Kalogeiton. 2022. Name your style: An arbitrary artist-aware image style transfer. arXiv:2202.13562. Retrieved from https:\/\/arxiv.org\/abs\/2202.13562"},{"key":"e_1_3_1_40_2","first-page":"5952","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","author":"Lu Ming","year":"2019","unstructured":"Ming Lu, Hao Zhao, Anbang Yao, Yurong Chen, Feng Xu, and Li Zhang. 2019. A closed-form solution to universal style transfer. In Proceedings of the IEEE\/CVF International Conference on Computer Vision. 5952\u20135961."},{"key":"e_1_3_1_41_2","unstructured":"Xuan Luo Zhen Han Lingkang Yang and Lingling Zhang. 2022. Consistent style transfer. arXiv:2201.02233. Retrieved from https:\/\/arxiv.org\/abs\/2201.02233"},{"key":"e_1_3_1_42_2","first-page":"331","volume-title":"Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision","author":"Ma Yingnan","year":"2023","unstructured":"Yingnan Ma, Chenqiu Zhao, Xudong Li, and Anup Basu. 2023. RAST: Restorable arbitrary style transfer via multi-restoration. In Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision. 331\u2013340."},{"key":"e_1_3_1_43_2","doi-asserted-by":"publisher","DOI":"10.1145\/3485061"},{"key":"e_1_3_1_44_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00603"},{"key":"e_1_3_1_45_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58583-9_45"},{"key":"e_1_3_1_46_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58545-7_19"},{"key":"e_1_3_1_47_2","first-page":"698","volume-title":"Proceedings of the European Conference on Computer Vision","author":"Sanakoyeu Artsiom","year":"2018","unstructured":"Artsiom Sanakoyeu, Dmytro Kotovenko, Sabine Lang, and Bjorn Ommer. 2018. A style-aware content loss for real-time HD style transfer. In Proceedings of the European Conference on Computer Vision. 698\u2013714."},{"key":"e_1_3_1_48_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00860"},{"key":"e_1_3_1_49_2","doi-asserted-by":"crossref","unstructured":"YiChang Shih Sylvain Paris Connelly Barnes William T. Freeman and Fr\u00e9do Durand. 2014. Style transfer for headshot portraits. ACM Transactions on Graphics (TOG) 33 4 (2014) 1\u201314.","DOI":"10.1145\/2601097.2601137"},{"key":"e_1_3_1_50_2","doi-asserted-by":"publisher","DOI":"10.1145\/3311781"},{"key":"e_1_3_1_51_2","unstructured":"Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networks for large-scale image recognition. 3rd International Conference on Learning Representations (ICLR\u201915)."},{"key":"e_1_3_1_52_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00357"},{"key":"e_1_3_1_53_2","first-page":"839","volume-title":"Proceedings of the 6th International Conference on Computer Vision","author":"Tomasi Carlo","year":"1998","unstructured":"Carlo Tomasi and Roberto Manduchi. 1998. Bilateral filtering for gray and color images. In Proceedings of the 6th International Conference on Computer Vision. IEEE, 839\u2013846."},{"key":"e_1_3_1_54_2","unstructured":"Dmitry Ulyanov Vadim Lebedev Andrea Vedaldi and Victor Lempitsky. 2016. Texture networks: Feed-forward synthesis of textures and stylized images. In Proceedings of the 33rd International Conference on International Conference on Machine Learning 48 (2016) 1349\u20131357."},{"key":"e_1_3_1_55_2","article-title":"Attention is all you need","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems.","journal-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems"},{"key":"e_1_3_1_56_2","doi-asserted-by":"publisher","DOI":"10.1145\/3506710"},{"key":"e_1_3_1_57_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00917"},{"key":"e_1_3_1_58_2","doi-asserted-by":"publisher","DOI":"10.1145\/1141911.1142018"},{"key":"e_1_3_1_59_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01041"},{"key":"e_1_3_1_60_2","first-page":"14618","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","author":"Wu Xiaolei","year":"2021","unstructured":"Xiaolei Wu, Zhihao Hu, Lu Sheng, and Dong Xu. 2021. Styleformer: Real-time arbitrary style transfer via parametric style composition. In Proceedings of the IEEE\/CVF International Conference on Computer Vision. 14618\u201314627."},{"key":"e_1_3_1_61_2","first-page":"7354","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Zhang Han","year":"2019","unstructured":"Han Zhang, Ian Goodfellow, Dimitris Metaxas, and Augustus Odena. 2019. Self-attention generative adversarial networks. In Proceedings of the International Conference on Machine Learning. PMLR, 7354\u20137363."},{"key":"e_1_3_1_62_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"e_1_3_1_63_2","doi-asserted-by":"crossref","unstructured":"Yuxin Zhang Fan Tang Weiming Dong Haibin Huang Chongyang Ma Tong-Yee Lee and Changsheng Xu. 2022. Domain enhanced arbitrary image style transfer via contrastive learning. ACM SIGGRAPH 2022 Conference Proceeding. 1\u20138.","DOI":"10.1145\/3528233.3530736"},{"issue":"2","key":"e_1_3_1_64_2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3542820","article-title":"Local correlation ensemble with GCN based on attention features for cross-domain person Re-ID","volume":"19","author":"Zhang Yue","year":"2023","unstructured":"Yue Zhang, Fanghui Zhang, Yi Jin, Yigang Cen, Viacheslav Voronin, and Shaohua Wan. 2023. Local correlation ensemble with GCN based on attention features for cross-domain person Re-ID. ACM Transactions on Multimedia Computing, Communications and Applications 19, 2 (2023), 1\u201322.","journal-title":"ACM Transactions on Multimedia Computing, Communications and Applications"},{"key":"e_1_3_1_65_2","volume-title":"Proceedings of the 2017 IEEE International Conference on Computer Vision","author":"Zhu Jun-Yan","year":"2017","unstructured":"Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A. Efros. 2017. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision."}],"container-title":["ACM Transactions on Multimedia Computing, Communications, and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3638770","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3638770","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:03:34Z","timestamp":1750291414000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3638770"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,22]]},"references-count":64,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,5,31]]}},"alternative-id":["10.1145\/3638770"],"URL":"https:\/\/doi.org\/10.1145\/3638770","relation":{},"ISSN":["1551-6857","1551-6865"],"issn-type":[{"type":"print","value":"1551-6857"},{"type":"electronic","value":"1551-6865"}],"subject":[],"published":{"date-parts":[[2024,1,22]]},"assertion":[{"value":"2023-06-05","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-12-21","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-01-22","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}