{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,22]],"date-time":"2025-05-22T06:26:28Z","timestamp":1747895188203,"version":"3.38.0"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,6,21]],"date-time":"2024-06-21T00:00:00Z","timestamp":1718928000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,6,21]],"date-time":"2024-06-21T00:00:00Z","timestamp":1718928000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Science and Technology Planning Project of Henan Province","award":["242102211003"],"award-info":[{"award-number":["242102211003"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62302297","61972157"],"award-info":[{"award-number":["62302297","61972157"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,3]]},"DOI":"10.1007\/s00371-024-03550-9","type":"journal-article","created":{"date-parts":[[2024,6,21]],"date-time":"2024-06-21T14:02:57Z","timestamp":1718978577000},"page":"2529-2544","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["MCLGAN: a multi-style cartoonization method based on style condition information"],"prefix":"10.1007","volume":"41","author":[{"given":"Canlin","family":"Li","sequence":"first","affiliation":[]},{"given":"Xinyue","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Ran","family":"Yi","sequence":"additional","affiliation":[]},{"given":"Wenjiao","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Lihua","family":"Bi","sequence":"additional","affiliation":[]},{"given":"Lizhuang","family":"Ma","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,21]]},"reference":[{"key":"3550_CR1","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., WardeFarley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. Adv. Neural Inf. Process Syst. 27 (2014)"},{"key":"3550_CR2","doi-asserted-by":"crossref","unstructured":"Chen, Y. C., Xu, X., Jia, J.: Domain adaptive image-to-image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5274\u20135283 (2020)","DOI":"10.1109\/CVPR42600.2020.00532"},{"key":"3550_CR3","doi-asserted-by":"crossref","unstructured":"Wang, Y., Khan, S., Gonzalez-Garcia, A., et al.: Semi-supervised learning for few-shot image-to-image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4453\u20134462(2020)","DOI":"10.1109\/CVPR42600.2020.00451"},{"key":"3550_CR4","doi-asserted-by":"crossref","unstructured":"Zhang, P., Zhang, B., Chen, D., et al.: Cross-domain correspondence learning for exemplar-based image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5143\u20135153(2020)","DOI":"10.1109\/CVPR42600.2020.00519"},{"key":"3550_CR5","doi-asserted-by":"crossref","unstructured":"Lin, L., Zhang, S., Ji, S., et al.: TMGAN: two-stage multi-domain generative adversarial network for landscape image translation. Vis. Comput. 1\u201317 (2023)","DOI":"10.1007\/s00371-023-03171-8"},{"key":"3550_CR6","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1016\/j.neucom.2019.08.075","volume":"370","author":"H Wu","year":"2019","unstructured":"Wu, H., Sun, Z., Zhang, Y., et al.: Direction-aware neural style transfer with texture enhancement. Neurocomputing 370, 39\u201355 (2019)","journal-title":"Neurocomputing"},{"issue":"2","key":"3550_CR7","doi-asserted-by":"publisher","first-page":"609","DOI":"10.1007\/s00371-021-02361-6","volume":"39","author":"W Ye","year":"2023","unstructured":"Ye, W., Zhu, X., Liu, Y.: Multi-semantic preserving neural style transfer based on Y channel information of image. Vis. Comput. 39(2), 609\u2013623 (2023)","journal-title":"Vis. Comput."},{"key":"3550_CR8","doi-asserted-by":"crossref","unstructured":"Yu, X., Zhou, G.: Arbitrary style transfer via content consistency and style consistency. Vis. Comput. 1\u201314 (2023)","DOI":"10.1007\/s00371-023-02855-5"},{"key":"3550_CR9","doi-asserted-by":"crossref","unstructured":"Xu, L., Yuan, Q., Sun, Y., et al.: Image neural style transfer combining global and local optimization. Vis. Comput. 1\u201315 (2024)","DOI":"10.1007\/s00371-023-03244-8"},{"key":"3550_CR10","doi-asserted-by":"crossref","unstructured":"Chen, Y., Lai, Y. K., Liu, Y. J.: Cartoongan: Generative adversarial networks for photo cartoonization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9465\u20139474 (2018)","DOI":"10.1109\/CVPR.2018.00986"},{"issue":"10","key":"3550_CR11","doi-asserted-by":"publisher","first-page":"3376","DOI":"10.1109\/TVCG.2021.3067201","volume":"28","author":"Y Shu","year":"2021","unstructured":"Shu, Y., Yi, R., Xia, M., et al.: Gan-based multi-style photo cartoonization. IEEE Trans. Vis. Comput. Graphics. 28(10), 3376\u20133390 (2021)","journal-title":"IEEE Trans. Vis. Comput. Graphics."},{"key":"3550_CR12","unstructured":"Mei, H., Chen, Z. J.: Cartoonish rendering of images based on mean shift and FDoG. Comput. Eng. Appl. 52 (10): 213\u2013217 (2016) (in Chinese)"},{"key":"3550_CR13","unstructured":"Liu, X.: Image cartoon processing based on Mean Shift in OpencCV. Information and Computer (Theory Edition). 32(20): 54\u201357(2020)(in Chinese)"},{"issue":"5","key":"3550_CR14","doi-asserted-by":"publisher","first-page":"603","DOI":"10.1109\/34.1000236","volume":"24","author":"D Comaniciu","year":"2002","unstructured":"Comaniciu, D., Meer, P.: Mean shift: A robust approach toward feature space analysis. IEEE Trans. Pattern Anal. Mach. Intell. 24(5), 603\u2013619 (2002)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"17","key":"3550_CR15","doi-asserted-by":"publisher","first-page":"4624","DOI":"10.1016\/j.ijleo.2014.05.031","volume":"125","author":"R Xue","year":"2014","unstructured":"Xue, R., Song, H., Wu, Z., et al.: An extended flow-based difference-of-Gaussians method of line drawing for polyhedral image. Optik 125(17), 4624\u20134628 (2014)","journal-title":"Optik"},{"key":"3550_CR16","doi-asserted-by":"publisher","first-page":"679","DOI":"10.1109\/TPAMI.1986.4767851","volume":"6","author":"J Canny","year":"1986","unstructured":"Canny, J.: A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 6, 679\u2013698 (1986)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3550_CR17","doi-asserted-by":"crossref","unstructured":"Chen, Y., Lai, Y. K., Liu, Y. J.: Transforming photos to comics using convolutional neural networks. In:Proceedings of the IEEE\/CVF Conference on Image Processing, pp. 2010\u20132014 (2017)","DOI":"10.1109\/ICIP.2017.8296634"},{"key":"3550_CR18","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1016\/j.neunet.2020.08.011","volume":"132","author":"W Li","year":"2016","unstructured":"Li, W., Xiong, W., Liao, H., et al.: Carigan: Caricature generation through weakly paired adversarial learning. Neural Netw. 132, 66\u201374 (2016)","journal-title":"Neural Netw."},{"key":"3550_CR19","doi-asserted-by":"crossref","unstructured":"Chen, J., Liu, G., Chen, X.: Animegan: A novel lightweight GAN for photo animation. In: Artificial Intelligence Algorithms and Applications: 11th International Symposium, ISICA 2019, Guangzhou, China, November 16\u201317, 2019, Revised Selected Papers 11. Springer Singapore, pp. 242\u2013256 (2020)","DOI":"10.1007\/978-981-15-5577-0_18"},{"key":"3550_CR20","doi-asserted-by":"crossref","unstructured":"Gatys, L. A., Ecker, A. S., Bethge, M.: Image style transfer using convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2414\u20132423 (2016)","DOI":"10.1109\/CVPR.2016.265"},{"key":"3550_CR21","doi-asserted-by":"publisher","first-page":"374","DOI":"10.1109\/TIP.2020.3036754","volume":"30","author":"R Li","year":"2020","unstructured":"Li, R., Wu, C.H., Liu, S., et al.: SDP-GAN: Saliency detail preservation generative adversarial networks for high perceptual quality style transfer. IEEE Trans. Image Process. 30, 374\u2013385 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"3550_CR22","doi-asserted-by":"publisher","first-page":"485","DOI":"10.1109\/TIP.2021.3130539","volume":"31","author":"Y Dong","year":"2021","unstructured":"Dong, Y., Tan, W., Tao, D., et al.: CartoonLossGAN: Learning surface and coloring of images for cartoonization. IEEE Trans. Image Process. 31, 485\u2013498 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"3550_CR23","unstructured":"Gao X., Zhang, Y., Tian, Y.: Learning to incorporate texture saliency adaptive attention to image cartoonization. In: ICML, 2(3): 6 (2022)"},{"key":"3550_CR24","doi-asserted-by":"crossref","unstructured":"Chen, Y., Chen, M., Song, C., et al.: Cartoonrenderer: An instance-based multi-style cartoon image translator. In: MultiMedia Modeling: 26th International Conference, MMM 2020, Daejeon, South Korea, January 5-8, 2020, Proceedings, Part I 26. Springer International Publishing, pp. 176\u2013187 (2020)","DOI":"10.1007\/978-3-030-37731-1_15"},{"key":"3550_CR25","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, J.: Learning to cartoonize using white-box cartoon representations. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8090\u20138099 (2020)","DOI":"10.1109\/CVPR42600.2020.00811"},{"key":"3550_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2023.126654","volume":"556","author":"ZJ Zhang","year":"2023","unstructured":"Zhang, Z.J., Sun, J.K., Chen, J.F., et al.: Caster: cartoon style transfer via dynamic cartoon style casting. Neurocomputing 556, 126654 (2023)","journal-title":"Neurocomputing"},{"key":"3550_CR27","unstructured":"Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv:1411.1784 (2014)"},{"key":"3550_CR28","unstructured":"Odena, A., Olah, C., Shlens, J.: Conditional image synthesis with auxiliary classifier GANs. In: International Conference on Machine Learning. PMLR, pp. 2642\u20132651 (2017)"},{"key":"3550_CR29","unstructured":"Miyato T., Koyama M.: cGANs with projection discriminator. In: International Conference on Learning Representations, (2018)"},{"key":"3550_CR30","first-page":"21357","volume":"33","author":"M Kang","year":"2020","unstructured":"Kang, M., Park, J.: Contragan: contrastive learning for conditional image generation. Adv Neural Inf. Process Syst. 33, 21357\u201321369 (2020)","journal-title":"Adv Neural Inf. Process Syst."},{"key":"3550_CR31","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-assisted Intervention-MICCAI 2015: 18th International Conference, Munich, Germany, 2015, proceedings, part III 18. Springer International Publishing, pp. 234\u2013241 (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"3550_CR32","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. CoRR abs\/1409.1556 (2014)"},{"key":"3550_CR33","unstructured":"Dumoulin, V., Shlens, J., Kudlur, M.: A learned representation for artistic style. CoRR, abs\/1610.07629, (2016)"},{"key":"3550_CR34","unstructured":"Miyato, T., Kataoka, T., Koyama, M., et al.: Spectral normalization for generative adversarial networks. In: International Conference on Learning Representations (2018)"},{"key":"3550_CR35","doi-asserted-by":"crossref","unstructured":"Zhu, J. Y., Park, T., Isola, P., et al.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International conference on computer vision, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"},{"issue":"5","key":"3550_CR36","doi-asserted-by":"publisher","first-page":"2111","DOI":"10.1007\/s00371-022-02468-4","volume":"39","author":"J Rao","year":"2023","unstructured":"Rao, J., Ke, A., Liu, G., et al.: MS-GAN: multi-scale GAN with parallel class activation maps for image reconstruction. Visual Comput. 39(5), 2111\u20132126 (2023)","journal-title":"Visual Comput."},{"key":"3550_CR37","doi-asserted-by":"crossref","unstructured":"Chen, Z., Zhang, Y.: CA-GAN: the synthesis of Chinese art paintings using generative adversarial networks. Visual Comput. 1\u201313 (2023)","DOI":"10.1007\/s00371-023-03115-2"},{"key":"3550_CR38","unstructured":"Kingma, D. P., Ba, J.: Adam: A method for stochastic optimization. In: Yoshua Bengio and Yann LeCun, editors, 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 2015, Conference Track Proceedings (2015)"},{"key":"3550_CR39","doi-asserted-by":"crossref","unstructured":"Luo, X., Han, Z., Yang, L.: Progressive attentional manifold alignment for arbitrary style transfer. In: Proceedings of the Asian Conference on Computer Vision, pp. 3206\u20133222 (2022)","DOI":"10.1007\/978-3-031-26293-7_9"},{"key":"3550_CR40","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., et al.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. Adv. Neural Inf. Process Syst. 30 (2017)"},{"key":"3550_CR41","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., et al.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818\u20132826 (2016)","DOI":"10.1109\/CVPR.2016.308"},{"key":"3550_CR42","unstructured":"Sutherland, J. D., Arbel, M., Gretton, A.: Demystifying MMD GANs. In: International Conference for Learning Representations, pp. 1\u201336 (2018)"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03550-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-024-03550-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03550-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,3]],"date-time":"2025-03-03T11:30:59Z","timestamp":1741001459000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-024-03550-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,21]]},"references-count":42,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,3]]}},"alternative-id":["3550"],"URL":"https:\/\/doi.org\/10.1007\/s00371-024-03550-9","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2024,6,21]]},"assertion":[{"value":"11 June 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 June 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no Conflict of interest to this work.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This chapter does not contain any studies with human participants or animals performed by any of the authors.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}