{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,7]],"date-time":"2026-05-07T20:36:40Z","timestamp":1778186200257,"version":"3.51.4"},"reference-count":49,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2023,12,27]],"date-time":"2023-12-27T00:00:00Z","timestamp":1703635200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,27]],"date-time":"2023-12-27T00:00:00Z","timestamp":1703635200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the National Key Research and Development Program of China","award":["2021ZD0111405"],"award-info":[{"award-number":["2021ZD0111405"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2024,4]]},"DOI":"10.1007\/s00521-023-09345-8","type":"journal-article","created":{"date-parts":[[2023,12,27]],"date-time":"2023-12-27T21:28:33Z","timestamp":1703712513000},"page":"5267-5284","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":14,"title":["DLP-GAN: learning to draw modern Chinese landscape photos with generative adversarial network"],"prefix":"10.1007","volume":"36","author":[{"given":"Xiangquan","family":"Gui","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2958-7488","authenticated-orcid":false,"given":"Binxuan","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Li","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yi","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,27]]},"reference":[{"key":"9345_CR1","doi-asserted-by":"publisher","unstructured":"Liu L (2021) The basic features of traditional Chinese landscape painting. In: The 5th international conference on art studies: research, experience, education (ICASSEE 2021), vol. 1, pp 17\u201327 . https:\/\/doi.org\/10.5117\/9789048557240\/ICASSEE.2021.003. Amsterdam University Press","DOI":"10.5117\/9789048557240\/ICASSEE.2021.003"},{"key":"9345_CR2","unstructured":"Li Y, Fang C, Yang J, Wang Z, Lu X, Yang M-H (2017) Universal style transfer via feature transforms. Adv Neural Inf Process Syst 30"},{"key":"9345_CR3","doi-asserted-by":"publisher","unstructured":"Gatys LA, Ecker AS, Bethge M (2016) Image style transfer using convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2414\u20132423 . https:\/\/doi.org\/10.1109\/cvpr.2016.265","DOI":"10.1109\/cvpr.2016.265"},{"key":"9345_CR4","doi-asserted-by":"publisher","unstructured":"Johnson J, Alahi A, Fei-Fei L (2016) Perceptual losses for real-time style transfer and super-resolution. In: European conference on computer vision, pp 694\u2013711. https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43. Springer","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"9345_CR5","doi-asserted-by":"publisher","unstructured":"Zhu JY, Park T, Isola P, Efros AA (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2223\u20132232 . https:\/\/doi.org\/10.1109\/iccv.2017.244","DOI":"10.1109\/iccv.2017.244"},{"key":"9345_CR6","unstructured":"Zhu J-Y, Zhang R, Pathak D, Darrell T, Efros AA, Wang O, Shechtman E (2017) Toward multimodal image-to-image translation. Adv Neural Inf Process Syst 30"},{"key":"9345_CR7","doi-asserted-by":"publisher","unstructured":"Isola P, Zhu J-Y, Zhou T, Efros AA (2017) Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1125\u20131134 . https:\/\/doi.org\/10.1109\/cvpr.2017.632","DOI":"10.1109\/cvpr.2017.632"},{"key":"9345_CR8","doi-asserted-by":"publisher","first-page":"374","DOI":"10.1109\/TIP.2020.3036754","volume":"30","author":"R Li","year":"2020","unstructured":"Li R, Wu C-H, Liu S, Wang J, Wang G, Liu G, Zeng B (2020) Sdp-gan: saliency detail preservation generative adversarial networks for high perceptual quality style transfer. IEEE Trans Image Process 30:374\u2013385. https:\/\/doi.org\/10.1109\/TIP.2020.3036754","journal-title":"IEEE Trans Image Process"},{"key":"9345_CR9","doi-asserted-by":"publisher","unstructured":"Lin T, Ma Z, Li F, He D, Li X, Ding E, Wang N, Li J, Gao X (2021) Drafting and revision: Laplacian pyramid network for fast high-quality artistic style transfer. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5141\u20135150 . https:\/\/doi.org\/10.1109\/cvpr46437.2021.00510","DOI":"10.1109\/cvpr46437.2021.00510"},{"key":"9345_CR10","doi-asserted-by":"publisher","unstructured":"Liu S, Lin T, He D, Li F, Wang M, Li X, Sun Z, Li Q, Ding E (2021) Adaattn: revisit attention mechanism in arbitrary neural style transfer. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 6649\u20136658 . https:\/\/doi.org\/10.1109\/iccv48922.2021.00658","DOI":"10.1109\/iccv48922.2021.00658"},{"key":"9345_CR11","doi-asserted-by":"publisher","unstructured":"Peng X, Peng S, Hu Q, Peng J, Wang J, Liu X, Fan J (2022) Contour-enhanced cyclegan framework for style transfer from scenery photos to Chinese landscape paintings. Neural Comput Appl 1\u201322 (2022). https:\/\/doi.org\/10.1007\/s00521-022-07432-w","DOI":"10.1007\/s00521-022-07432-w"},{"key":"9345_CR12","doi-asserted-by":"publisher","unstructured":"Zheng C, Zhang Y (2018) Two-stage color ink painting style transfer via convolution neural network. In: 2018 15th international symposium on pervasive systems, algorithms and networks (I-SPAN), pp 193\u2013200. https:\/\/doi.org\/10.1109\/i-span.2018.00039. IEEE","DOI":"10.1109\/i-span.2018.00039"},{"key":"9345_CR13","doi-asserted-by":"publisher","unstructured":"Zhou L, Wang Q-F, Huang K, Lo C-H (2019) An interactive and generative approach for Chinese Shanshui painting document. In: 2019 International conference on document analysis and recognition (ICDAR), pp 819\u2013824. https:\/\/doi.org\/10.1109\/icdar.2019.00136. IEEE","DOI":"10.1109\/icdar.2019.00136"},{"issue":"11","key":"9345_CR14","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2020) Generative adversarial networks. Commun ACM 63(11):139\u2013144. https:\/\/doi.org\/10.1145\/3422622","journal-title":"Commun ACM"},{"issue":"24","key":"9345_CR15","doi-asserted-by":"publisher","first-page":"21433","DOI":"10.1007\/s00521-021-05975-y","volume":"34","author":"V Bharti","year":"2022","unstructured":"Bharti V, Biswas B, Shukla KK (2022) Emocgan: a novel evolutionary multiobjective cyclic generative adversarial network and its application to unpaired image translation. Neural Comput Appl 34(24):21433\u201321447. https:\/\/doi.org\/10.1007\/s00521-021-05975-y","journal-title":"Neural Comput Appl"},{"key":"9345_CR16","doi-asserted-by":"publisher","unstructured":"He B, Gao F, Ma D, Shi B, Duan L-Y (2018) Chipgan: a generative adversarial network for Chinese ink wash painting style transfer. In: Proceedings of the 26th ACM international conference on multimedia, pp 1172\u20131180. https:\/\/doi.org\/10.1145\/3240508.3240655","DOI":"10.1145\/3240508.3240655"},{"key":"9345_CR17","doi-asserted-by":"publisher","unstructured":"Wang W, Li Y, Ye H, Ye F, Xu X (2022) Ink painting style transfer using asymmetric cycle-consistent GAN. Available at SSRN 4109972 . https:\/\/doi.org\/10.2139\/ssrn.4109972","DOI":"10.2139\/ssrn.4109972"},{"key":"9345_CR18","doi-asserted-by":"publisher","unstructured":"Li B, Xiong C, Wu T, Zhou Y, Zhang L, Chu R (2018) Neural abstract style transfer for Chinese traditional painting. In: Asian conference on computer vision, pp 212\u2013227 . https:\/\/doi.org\/10.1007\/978-3-030-20890-5_14. Springer","DOI":"10.1007\/978-3-030-20890-5_14"},{"key":"9345_CR19","doi-asserted-by":"publisher","unstructured":"Qiao T, Zhang W, Zhang M, Ma Z, Xu D (2019) Ancient painting to natural image: a new solution for painting processing. In: 2019 IEEE winter conference on applications of computer vision (WACV), pp 521\u2013530. https:\/\/doi.org\/10.1109\/wacv.2019.00061","DOI":"10.1109\/wacv.2019.00061"},{"issue":"24","key":"9345_CR20","doi-asserted-by":"publisher","first-page":"21551","DOI":"10.1007\/s00521-021-06147-8","volume":"34","author":"S Qin","year":"2022","unstructured":"Qin S, Liu S (2022) Towards end-to-end car license plate location and recognition in unconstrained scenarios. Neural Comput Appl 34(24):21551\u201321566. https:\/\/doi.org\/10.1007\/s00521-021-06147-8","journal-title":"Neural Comput Appl"},{"key":"9345_CR21","doi-asserted-by":"publisher","unstructured":"Sun H, Wu L, Li X, Meng X (2022) Style-woven attention network for zero-shot ink wash painting style transfer. In: Proceedings of the 2022 international conference on multimedia retrieval, pp 277\u2013285. https:\/\/doi.org\/10.1145\/3512527.3531391","DOI":"10.1145\/3512527.3531391"},{"key":"9345_CR22","doi-asserted-by":"publisher","unstructured":"Li J, Wang Q, Li S, Zhong Q, Zhou Q (2021) Immersive traditional Chinese portrait painting: research on style transfer and face replacement. In: Chinese conference on pattern recognition and computer vision (PRCV), pp 192\u2013203. https:\/\/doi.org\/10.1007\/978-3-030-88007-1_16. Springer","DOI":"10.1007\/978-3-030-88007-1_16"},{"key":"9345_CR23","doi-asserted-by":"publisher","unstructured":"Xue A (2021) End-to-end Chinese landscape painting creation using generative adversarial networks. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 3863\u20133871. https:\/\/doi.org\/10.1109\/wacv48630.2021.00391","DOI":"10.1109\/wacv48630.2021.00391"},{"key":"9345_CR24","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal P, Nichol A (2021) Diffusion models beat GANs on image synthesis. Adv Neural Inf Process Syst 34:8780\u20138794","journal-title":"Adv Neural Inf Process Syst"},{"key":"9345_CR25","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho J, Jain A, Abbeel P (2020) Denoising diffusion probabilistic models. Adv Neural Inf Process Syst 33:6840\u20136851","journal-title":"Adv Neural Inf Process Syst"},{"key":"9345_CR26","doi-asserted-by":"publisher","unstructured":"Saharia C, Chan W, Chang H, Lee C, Ho J, Salimans T, Fleet D, Norouzi M (2022) Palette: image-to-image diffusion models. In: ACM SIGGRAPH 2022 conference proceedings, pp 1\u201310. https:\/\/doi.org\/10.1145\/3528233.3530757","DOI":"10.1145\/3528233.3530757"},{"key":"9345_CR27","doi-asserted-by":"publisher","unstructured":"Su X, Song J, Meng C, Ermon S (2022) Dual diffusion implicit bridges for image-to-image translation. arXiv preprint arXiv:2203.08382. https:\/\/doi.org\/10.48550\/arXiv.2203.08382","DOI":"10.48550\/arXiv.2203.08382"},{"key":"9345_CR28","doi-asserted-by":"crossref","unstructured":"Li B, Xue K, Liu B, Lai Y-K (2023) Bbdm: image-to-image translation with brownian bridge diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern Recognition, pp 1952\u20131961","DOI":"10.1109\/CVPR52729.2023.00194"},{"issue":"5","key":"9345_CR29","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/tip.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li H, Wu X-J (2018) Densefuse: a fusion approach to infrared and visible images. IEEE Trans Image Process 28(5):2614\u20132623. https:\/\/doi.org\/10.1109\/tip.2018.2887342","journal-title":"IEEE Trans Image Process"},{"key":"9345_CR30","doi-asserted-by":"publisher","unstructured":"Wang T-C, Liu M-Y, Zhu J-Y, Tao A, Kautz J, Catanzaro B (2018) High-resolution image synthesis and semantic manipulation with conditional GANs. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8798\u20138807. https:\/\/doi.org\/10.1109\/cvpr.2018.00917","DOI":"10.1109\/cvpr.2018.00917"},{"key":"9345_CR31","doi-asserted-by":"publisher","unstructured":"Huang X, Liu M-Y, Belongie S, Kautz J (2018) Multimodal unsupervised image-to-image translation. In: Proceedings of the European conference on computer vision (ECCV), pp 172\u2013189. https:\/\/doi.org\/10.1007\/978-3-030-01219-9_11","DOI":"10.1007\/978-3-030-01219-9_11"},{"key":"9345_CR32","doi-asserted-by":"publisher","first-page":"132002","DOI":"10.1109\/access.2020.3009470","volume":"8","author":"F Zhang","year":"2020","unstructured":"Zhang F, Gao H, Lai Y (2020) Detail-preserving cyclegan-adain framework for image-to-ink painting translation. IEEE Access 8:132002\u2013132011. https:\/\/doi.org\/10.1109\/access.2020.3009470","journal-title":"IEEE Access"},{"key":"9345_CR33","doi-asserted-by":"publisher","unstructured":"Chung C-Y, Huang S-H (2022) Interactively transforming chinese ink paintings into realistic images using a border enhance generative adversarial network. Multimedia Tools Appl 1\u201334. https:\/\/doi.org\/10.1007\/s11042-022-13684-4","DOI":"10.1007\/s11042-022-13684-4"},{"key":"9345_CR34","doi-asserted-by":"publisher","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778. https:\/\/doi.org\/10.1109\/cvpr.2016.90","DOI":"10.1109\/cvpr.2016.90"},{"key":"9345_CR35","doi-asserted-by":"publisher","unstructured":"Huang G, Liu Z, Van Der\u00a0Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4700\u20134708. https:\/\/doi.org\/10.1109\/cvpr.2017.243","DOI":"10.1109\/cvpr.2017.243"},{"key":"9345_CR36","doi-asserted-by":"publisher","unstructured":"Mao X, Li Q, Xie H, Lau RY, Wang Z, Paul\u00a0Smolley S (2017) Least squares generative adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2794\u20132802. https:\/\/doi.org\/10.1109\/iccv.2017.304","DOI":"10.1109\/iccv.2017.304"},{"key":"9345_CR37","doi-asserted-by":"publisher","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556. https:\/\/doi.org\/10.48550\/arXiv.1409.1556","DOI":"10.48550\/arXiv.1409.1556"},{"key":"9345_CR38","doi-asserted-by":"publisher","unstructured":"Poma XS, Riba E, Sappa A (2020) Dense extreme inception network: towards a robust CNN model for edge detection. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 1923\u20131932. https:\/\/doi.org\/10.1109\/wacv45572.2020.9093290","DOI":"10.1109\/wacv45572.2020.9093290"},{"key":"9345_CR39","doi-asserted-by":"publisher","unstructured":"Zhang R, Isola P, Efros AA, Shechtman E, Wang O (2018) The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 586\u2013595. https:\/\/doi.org\/10.1109\/cvpr.2018.00068","DOI":"10.1109\/cvpr.2018.00068"},{"key":"9345_CR40","unstructured":"Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, Killeen T, Lin Z, Gimelshein N, Antiga L (2019) Pytorch: an imperative style, high-performance deep learning library. Adv Neural Inf Process Syst 32"},{"key":"9345_CR41","doi-asserted-by":"publisher","unstructured":"Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980. https:\/\/doi.org\/10.48550\/arXiv.1412.6980","DOI":"10.48550\/arXiv.1412.6980"},{"key":"9345_CR42","doi-asserted-by":"publisher","unstructured":"Huang X, Belongie S (2017) Arbitrary style transfer in real-time with adaptive instance normalization. In: Proceedings of the IEEE international conference on computer vision, pp 1501\u20131510. https:\/\/doi.org\/10.1109\/iccv.2017.167","DOI":"10.1109\/iccv.2017.167"},{"key":"9345_CR43","doi-asserted-by":"publisher","first-page":"114","DOI":"10.1016\/j.neucom.2020.07.044","volume":"415","author":"H Dou","year":"2020","unstructured":"Dou H, Chen C, Hu X, Jia L, Peng S (2020) Asymmetric cyclegan for image-to-image translations with uneven complexities. Neurocomputing 415:114\u2013122. https:\/\/doi.org\/10.1016\/j.neucom.2020.07.044","journal-title":"Neurocomputing"},{"issue":"3","key":"9345_CR44","doi-asserted-by":"publisher","first-page":"619","DOI":"10.1007\/s41095-022-0295-3","volume":"9","author":"Z Peng","year":"2023","unstructured":"Peng Z, Wang H, Weng Y, Yang Y, Shao T (2023) Unsupervised image translation with distributional semantics awareness. Comput Vis Media 9(3):619\u2013631. https:\/\/doi.org\/10.1007\/s41095-022-0295-3","journal-title":"Comput Vis Media"},{"key":"9345_CR45","unstructured":"Liu M-Y, Breuel T, Kautz J (2017) Unsupervised image-to-image translation networks. Adv Neural Inf Process Syst 30"},{"key":"9345_CR46","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2021.3105725","author":"H Tang","year":"2021","unstructured":"Tang H, Liu H, Xu D, Torr PH, Sebe N (2021) Attentiongan: unpaired image-to-image translation using attention-guided generative adversarial networks. IEEE Trans Neural Netw Learn Syst. https:\/\/doi.org\/10.1109\/TNNLS.2021.3105725","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"9345_CR47","unstructured":"Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S (2017) Gans trained by a two time-scale update rule converge to a local nash equilibrium. Adv Neural Inf Process Syst 30"},{"key":"9345_CR48","doi-asserted-by":"publisher","unstructured":"Bi\u0144kowski M, Sutherland DJ, Arbel M, Gretton A (2018) Demystifying MMD GANs. arXiv preprint arXiv:1801.01401. https:\/\/doi.org\/10.48550\/arXiv.1801.01401","DOI":"10.48550\/arXiv.1801.01401"},{"key":"9345_CR49","doi-asserted-by":"publisher","unstructured":"Hore A, Ziou D (2010) Image quality metrics: Psnr vs. ssim. In: 2010 20th international conference on pattern recognition, pp 2366\u20132369. https:\/\/doi.org\/10.1109\/icpr.2010.579. IEEE","DOI":"10.1109\/icpr.2010.579"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-023-09345-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-023-09345-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-023-09345-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,8]],"date-time":"2024-03-08T21:55:56Z","timestamp":1709934956000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-023-09345-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,27]]},"references-count":49,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["9345"],"URL":"https:\/\/doi.org\/10.1007\/s00521-023-09345-8","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"value":"0941-0643","type":"print"},{"value":"1433-3058","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12,27]]},"assertion":[{"value":"17 January 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 November 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 December 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest to this work.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}