{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T03:05:03Z","timestamp":1740107103197,"version":"3.37.3"},"reference-count":45,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2024,1,24]],"date-time":"2024-01-24T00:00:00Z","timestamp":1706054400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,24]],"date-time":"2024-01-24T00:00:00Z","timestamp":1706054400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["(Project No.52065010 and No.52165063)"],"award-info":[{"award-number":["(Project No.52065010 and No.52165063)"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100004001","name":"Department of Science and Technology of Guizhou Province","doi-asserted-by":"crossref","award":["(Project No. [2022] G140; No. [2022] K024; [2023] G094; [2023] G125)"],"award-info":[{"award-number":["(Project No. [2022] G140; No. [2022] K024; [2023] G094; [2023] G125)"]}],"id":[{"id":"10.13039\/501100004001","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Laboratory Open Project of Guizhou University","award":["(SYSKF2023-089)"],"award-info":[{"award-number":["(SYSKF2023-089)"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1007\/s00371-023-03244-8","type":"journal-article","created":{"date-parts":[[2024,1,24]],"date-time":"2024-01-24T18:01:54Z","timestamp":1706119314000},"page":"8397-8411","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Image neural style transfer combining global and local optimization"],"prefix":"10.1007","volume":"40","author":[{"given":"Liangyao","family":"Xu","sequence":"first","affiliation":[]},{"given":"Qingni","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Yu","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Qingyang","family":"Gao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,24]]},"reference":[{"key":"3244_CR1","doi-asserted-by":"publisher","first-page":"108459","DOI":"10.1016\/j.compeleceng.2022.108459","volume":"104","author":"F Zhang","year":"2022","unstructured":"Zhang, F., Liang, X., Sun, Y., et al.: POFMakeup: A style transfer method for peking opera makeup. Comput. Electr. Eng. 104, 108459 (2022)","journal-title":"Comput. Electr. Eng."},{"key":"3244_CR2","first-page":"262","volume":"28","author":"L Gatys","year":"2015","unstructured":"Gatys, L., Ecker, A.S., Bethge, M.: Texture synthesis using convolutional neural networks. Adv. Neural. Inf. Process. Syst. 28, 262\u2013270 (2015)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"3244_CR3","doi-asserted-by":"crossref","unstructured":"Gatys, L.A., Ecker, A.S., Bethge, M.: A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576 (2015)","DOI":"10.1167\/16.12.326"},{"key":"3244_CR4","doi-asserted-by":"crossref","unstructured":"Gatys, L.A., Ecker, A.S., Bethge, M.: Image style transfer using convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, 2414\u20132423. (2016)","DOI":"10.1109\/CVPR.2016.265"},{"key":"3244_CR5","doi-asserted-by":"publisher","DOI":"10.1155\/2022\/8918722","author":"S Dong","year":"2022","unstructured":"Dong, S., Ding, Y., Qian, Y., et al.: Video style transfer based on convolutional neural networks. Math. Probl. Eng. (2022). https:\/\/doi.org\/10.1155\/2022\/8918722","journal-title":"Math. Probl. Eng."},{"key":"3244_CR6","doi-asserted-by":"publisher","first-page":"1911","DOI":"10.1109\/TIP.2022.3149237","volume":"31","author":"Y-S Liao","year":"2022","unstructured":"Liao, Y.-S., Huang, C.-R.: Semantic context-aware image style transfer. IEEE Trans. Image Process. 31, 1911\u20131923 (2022)","journal-title":"IEEE Trans. Image Process."},{"key":"3244_CR7","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-023-02915-w","author":"Y Li","year":"2023","unstructured":"Li, Y., Lin, G., He, M., et al.: Layer similarity guiding few-shot Chinese style transfer. Vis. Comput. (2023). https:\/\/doi.org\/10.1007\/s00371-023-02915-w","journal-title":"Vis. Comput."},{"key":"3244_CR8","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-022-02719-4","author":"J Lan","year":"2022","unstructured":"Lan, J., Ye, F., Ye, Z., et al.: Unsupervised style-guided cross-domain adaptation for few-shot stylized face translation. Vis. Comput. (2022). https:\/\/doi.org\/10.1007\/s00371-022-02719-4","journal-title":"Vis. Comput."},{"key":"3244_CR9","doi-asserted-by":"publisher","DOI":"10.1155\/2022\/5087129","author":"X Jin","year":"2022","unstructured":"Jin, X.: Art style transfer of oil painting based on parallel convolutional neural network. Secur. Commun. Netw. (2022). https:\/\/doi.org\/10.1155\/2022\/5087129","journal-title":"Secur. Commun. Netw."},{"key":"3244_CR10","doi-asserted-by":"publisher","first-page":"1491","DOI":"10.1080\/17538947.2023.2202422","volume":"16","author":"S Hong","year":"2023","unstructured":"Hong, S., Shen, J., L\u00fc, G., et al.: Aesthetic style transferring method based on deep neural network between Chinese landscape painting and classical private garden\u2019s virtual scenario. Int. J. Dig Earth 16, 1491\u20131509 (2023). https:\/\/doi.org\/10.1080\/17538947.2023.2202422","journal-title":"Int. J. Dig Earth"},{"key":"3244_CR11","doi-asserted-by":"crossref","unstructured":"Tang, Z., Wu, C., Xiao, Y., et al.: Evaluation of painting artistic style transfer based on generative adversarial network. In: 2023 8th International Conference on Cloud Computing and Big Data Analytics (ICCCBDA), 560\u2013566. IEEE, (2023)","DOI":"10.1109\/ICCCBDA56900.2023.10154714"},{"key":"3244_CR12","doi-asserted-by":"publisher","unstructured":"Luan, F., Paris, S., Shechtman, E., et al.: Deep photo style transfer. In: Proceedings of the IEEE Conference on Computer vision and pattern recognition, 4990\u20134998. (2017). https:\/\/doi.org\/10.48550\/arXiv.1703.07511","DOI":"10.48550\/arXiv.1703.07511"},{"key":"3244_CR13","doi-asserted-by":"crossref","unstructured":"Yang, R.R.: Multi-stage optimization for photorealistic neural style transfer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, 1769\u20131776, (2019)","DOI":"10.1109\/CVPRW.2019.00227"},{"key":"3244_CR14","doi-asserted-by":"crossref","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Computer Vision\u2013ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, 694\u2013711. Springer, (2016)","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"3244_CR15","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.-Y., Zhou, T., et al.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, 1125\u20131134. (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"3244_CR16","doi-asserted-by":"crossref","unstructured":"Li, C., Wand, M.: Combining markov random fields and convolutional neural networks for image synthesis. In: Proceedings of the IEEE conference on computer vision and pattern recognition, 2479\u20132486. (2016)","DOI":"10.1109\/CVPR.2016.272"},{"key":"3244_CR17","doi-asserted-by":"crossref","unstructured":"Kwatra, V., Essa, I., Bobick, A., et al.: Texture optimization for example-based synthesis. ACM SIGGRAPH 2005 Papers, 795\u2013802 (2005)","DOI":"10.1145\/1186822.1073263"},{"key":"3244_CR18","doi-asserted-by":"publisher","unstructured":"Li, S., Xu, X., Nie, L., et al.: Laplacian-steered neural style transfer. In: Proceedings of the 25th ACM International Conference on Multimedia, 1716\u20131724. (2017) https:\/\/doi.org\/10.1145\/3123266.3123425","DOI":"10.1145\/3123266.3123425"},{"key":"3244_CR19","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511996504","volume-title":"Computer Vision: Models, Learning, and Inference","author":"SJ Prince","year":"2012","unstructured":"Prince, S.J.: Computer Vision: Models, Learning, and Inference. Cambridge University Press, Cambridge (2012)"},{"key":"3244_CR20","doi-asserted-by":"publisher","first-page":"609","DOI":"10.1007\/s00371-021-02361-6","volume":"39","author":"W Ye","year":"2023","unstructured":"Ye, W., Zhu, X., Liu, Y.: Multi-semantic preserving neural style transfer based on Y channel information of image. Vis. Comput. 39, 609\u2013623 (2023). https:\/\/doi.org\/10.1007\/s00371-021-02361-6","journal-title":"Vis. Comput."},{"key":"3244_CR21","doi-asserted-by":"crossref","unstructured":"Ye, H., Liu, W., Huang, S.: Method of Image Style Transfer Based on Edge Detection. In: 2020 IEEE 4th Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), 1635\u20131639. IEEE, (2020)","DOI":"10.1109\/ITNEC48623.2020.9085127"},{"key":"3244_CR22","doi-asserted-by":"crossref","unstructured":"Ye, H., Xue, L., Chen, X., et al.: Research on the Method of Landscape Image Style Transfer based on Semantic Segmentation. In: 2021 IEEE 2nd International Conference on Information Technology, Big Data and Artificial Intelligence (ICIBA), 1171\u20131175. (2021)","DOI":"10.1109\/ICIBA52610.2021.9687957"},{"key":"3244_CR23","doi-asserted-by":"crossref","unstructured":"Zhu, J.-Y., Park, T., Isola, P., et al.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision, 2223\u20132232. (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"3244_CR24","doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M., Kim, M., et al.: Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, 8789\u20138797. (2018). https:\/\/arxiv.org\/abs\/1701.08893","DOI":"10.1109\/CVPR.2018.00916"},{"key":"3244_CR25","doi-asserted-by":"crossref","unstructured":"Zhu, S., Luo, X., Ma, L., et al.: Realistic Style-Transfer Generative Adversarial Network With a Weight-Sharing Strategy. In: 2020 IEEE 32nd International Conference on Tools with Artificial Intelligence (ICTAI), 694\u2013699. IEEE, (2020)","DOI":"10.1109\/ICTAI50040.2020.00111"},{"key":"3244_CR26","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-023-03042-2","author":"Y Yu","year":"2023","unstructured":"Yu, Y., Li, D., Li, B., et al.: Multi-style image generation based on semantic image. Vis. Comput. (2023). https:\/\/doi.org\/10.1007\/s00371-023-03042-2","journal-title":"Vis. Comput."},{"key":"3244_CR27","doi-asserted-by":"publisher","first-page":"3299","DOI":"10.1007\/s00371-023-02941-8","volume":"39","author":"Y Liu","year":"2023","unstructured":"Liu, Y., Guo, Z., Guo, H., et al.: Zoom-GAN: learn to colorize multi-scale targets. Vis. Comput. 39, 3299\u20133310 (2023). https:\/\/doi.org\/10.1007\/s00371-023-02941-8","journal-title":"Vis. Comput."},{"key":"3244_CR28","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-023-02814-0","author":"F Wang","year":"2023","unstructured":"Wang, F., Geng, S., Zhang, D., et al.: Automatic colorization for Thangka sketch-based paintings. Vis. Comput. (2023). https:\/\/doi.org\/10.1007\/s00371-023-02814-0","journal-title":"Vis. Comput."},{"key":"3244_CR29","doi-asserted-by":"publisher","first-page":"8526","DOI":"10.1109\/TIP.2021.3117061","volume":"30","author":"H Li","year":"2021","unstructured":"Li, H., Sheng, B., Li, P., et al.: Globally and locally semantic colorization via exemplar-based broad-GAN. IEEE Trans. Image Process. 30, 8526\u20138539 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"3244_CR30","doi-asserted-by":"crossref","unstructured":"Park, D.Y., Lee, K.H.: Arbitrary style transfer with style-attentional networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 5880\u20135888 (2019)","DOI":"10.1109\/CVPR.2019.00603"},{"key":"3244_CR31","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Fang, C., Wang, Y., et al.: Multimodal style transfer via graph cuts. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 5943\u20135951. (2019)","DOI":"10.1109\/ICCV.2019.00604"},{"key":"3244_CR32","doi-asserted-by":"publisher","unstructured":"Wang, H., Li, Y., Wang, Y., et al.: Collaborative distillation for ultra-resolution universal style transfer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 1860\u20131869 (2020). https:\/\/doi.org\/10.48550\/arXiv.2003.08436","DOI":"10.48550\/arXiv.2003.08436"},{"key":"3244_CR33","doi-asserted-by":"publisher","unstructured":"Li, Y., Fang, C., Yang, J., et al.: Diversified texture synthesis with feed-forward networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, 3920\u20133928. (2017). https:\/\/doi.org\/10.48550\/arXiv.1703.01664","DOI":"10.48550\/arXiv.1703.01664"},{"key":"3244_CR34","doi-asserted-by":"publisher","first-page":"546","DOI":"10.1109\/TIP.2018.2869695","volume":"28","author":"X Chen","year":"2018","unstructured":"Chen, X., Xu, C., Yang, X., et al.: Gated-gan: adversarial gated networks for multi-collection style transfer. IEEE Trans. Image Process. 28, 546\u2013560 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"3244_CR35","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11042-021-11694-2","volume":"81","author":"J Yu","year":"2022","unstructured":"Yu, J., Jin, L., Chen, J., et al.: Deep semantic space guided multi-scale neural style transfer. Multimed. Tools Appl. 81, 1\u201324 (2022)","journal-title":"Multimed. Tools Appl."},{"key":"3244_CR36","doi-asserted-by":"publisher","first-page":"135","DOI":"10.1016\/j.neucom.2020.04.027","volume":"406","author":"Z Ma","year":"2020","unstructured":"Ma, Z., Li, J., Wang, N., et al.: Semantic-related image style transfer with dual-consistency loss. Neurocomputing 406, 135\u2013149 (2020). https:\/\/doi.org\/10.1016\/j.neucom.2020.04.027","journal-title":"Neurocomputing"},{"key":"3244_CR37","doi-asserted-by":"publisher","first-page":"7930","DOI":"10.1109\/ACCESS.2021.3049637","volume":"9","author":"S Kim","year":"2021","unstructured":"Kim, S., Do, J., Kim, M.: Pseudo-supervised learning for semantic multi-style transfer. IEEE Access 9, 7930\u20137942 (2021)","journal-title":"IEEE Access"},{"key":"3244_CR38","doi-asserted-by":"crossref","unstructured":"Ma, Z., Lin, T., Li, X., et al.: Dual-affinity style embedding network for semantic-aligned image style transfer. IEEE Trans. Neural Networks Learn. Syst. (2022)","DOI":"10.1109\/TNNLS.2022.3143356"},{"key":"3244_CR39","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2021.103378","volume":"82","author":"W Ye","year":"2022","unstructured":"Ye, W., Zhu, X., Xu, Z., et al.: A comprehensive framework of multiple semantics preservation in neural style transfer. J. Vis. Commun. Image Represent. 82, 103378 (2022). https:\/\/doi.org\/10.1016\/j.jvcir.2021.103378","journal-title":"J. Vis. Commun. Image Represent."},{"key":"3244_CR40","unstructured":"Champandard, A.J.: Semantic style transfer and turning two-bit doodles into fine artworks. arXiv preprint arXiv:1603.01768 (2016)"},{"key":"3244_CR41","doi-asserted-by":"publisher","first-page":"575","DOI":"10.1049\/iet-cvi.2019.0844","volume":"14","author":"Z Wang","year":"2020","unstructured":"Wang, Z., Zhao, L., Lin, S., et al.: GLStyleNet: exquisite style transfer combining global and local pyramid features. IET Comput. Vis. 14, 575\u2013586 (2020). https:\/\/doi.org\/10.1049\/iet-cvi.2019.0844","journal-title":"IET Comput. Vis."},{"key":"3244_CR42","doi-asserted-by":"crossref","unstructured":"Li, C., Wand, M.: Precomputed real-time texture synthesis with markovian generative adversarial networks. In: Computer Vision\u2013ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 702\u2013716. (2016)","DOI":"10.1007\/978-3-319-46487-9_43"},{"key":"3244_CR43","doi-asserted-by":"publisher","first-page":"1307","DOI":"10.1007\/s00371-019-01726-2","volume":"36","author":"H-H Zhao","year":"2020","unstructured":"Zhao, H.-H., Rosin, P.L., Lai, Y.-K., et al.: Automatic semantic style transfer using deep convolutional neural networks and soft masks. Vis. Comput. 36, 1307\u20131324 (2020)","journal-title":"Vis. Comput."},{"key":"3244_CR44","doi-asserted-by":"publisher","first-page":"54518","DOI":"10.1109\/ACCESS.2021.3054969","volume":"9","author":"Z Lin","year":"2021","unstructured":"Lin, Z., Wang, Z., Chen, H., et al.: Image style transfer algorithm based on semantic segmentation. IEEE Access 9, 54518\u201354529 (2021)","journal-title":"IEEE Access"},{"key":"3244_CR45","doi-asserted-by":"crossref","unstructured":"Wu, C., Yao, M.: Automatically Extract Semantic Map for Semantic Style Transfer. In: 2022 7th International Conference on Intelligent Computing and Signal Processing (ICSP), 1127\u20131130. IEEE, (2022)","DOI":"10.1109\/ICSP54964.2022.9778781"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03244-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-03244-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03244-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,12]],"date-time":"2024-11-12T09:06:22Z","timestamp":1731402382000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-03244-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,24]]},"references-count":45,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["3244"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-03244-8","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2024,1,24]]},"assertion":[{"value":"19 December 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 January 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}