{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T07:20:57Z","timestamp":1765610457484},"reference-count":49,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2023,12,7]],"date-time":"2023-12-07T00:00:00Z","timestamp":1701907200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,7]],"date-time":"2023-12-07T00:00:00Z","timestamp":1701907200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,9]]},"DOI":"10.1007\/s00371-023-03171-8","type":"journal-article","created":{"date-parts":[[2023,12,7]],"date-time":"2023-12-07T19:02:01Z","timestamp":1701975721000},"page":"6389-6405","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["TMGAN: two-stage multi-domain generative adversarial network for landscape image translation"],"prefix":"10.1007","volume":"40","author":[{"given":"Liyuan","family":"Lin","sequence":"first","affiliation":[]},{"given":"Shun","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Shulin","family":"Ji","sequence":"additional","affiliation":[]},{"given":"Shuxian","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Aolin","family":"Wen","sequence":"additional","affiliation":[]},{"given":"Jingpeng","family":"Yan","sequence":"additional","affiliation":[]},{"given":"Yuan","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Weibin","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,7]]},"reference":[{"key":"3171_CR1","doi-asserted-by":"publisher","first-page":"108998","DOI":"10.1016\/j.patcog.2022.108998","volume":"133","author":"B Bosquet","year":"2023","unstructured":"Bosquet, B., Cores, D., Seidenari, L., Brea, V.M., Mucientes, M., Bimbo, A.D.: A full data augmentation pipeline for small object detection based on generative adversarial networks. Pattern Recognit. 133, 108998 (2023)","journal-title":"Pattern Recognit."},{"key":"3171_CR2","doi-asserted-by":"crossref","unstructured":"Chen, W., Hays, J.: Sketchygan: towards diverse and realistic sketch to image synthesis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9416\u20139425 (2018)","DOI":"10.1109\/CVPR.2018.00981"},{"key":"3171_CR3","doi-asserted-by":"crossref","unstructured":"Choi, Y., Uh, Y., Yoo, J., Ha, J.-W.: Stargan v2: diverse image synthesis for multiple domains. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 8188\u20138197 (2020)","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"3171_CR4","doi-asserted-by":"crossref","unstructured":"Deng, Y., Tang, F., Dong, W., Ma, C., Pan, X., Wang, L., Xu, C.: Stytr2: image style transfer with transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11326\u201311336 (2022)","DOI":"10.1109\/CVPR52688.2022.01104"},{"key":"3171_CR5","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et\u00a0al.: An image is worth 16x16 words: transformers for image recognition at scale (2020). arXiv:2010.11929"},{"key":"3171_CR6","doi-asserted-by":"crossref","unstructured":"Du, Z., Gao, L., Li, X.: A new contrastive gan with data augmentation for surface defect recognition under limited data. IEEE Trans. Instrum. Measure. (2022)","DOI":"10.1109\/TIM.2022.3232649"},{"key":"3171_CR7","doi-asserted-by":"crossref","unstructured":"Gao, C., Liu, Q., Xu, Q., Wang, L., Liu, J., Zou, C.: Sketchycoco: image generation from freehand scene sketches. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5174\u20135183 (2020)","DOI":"10.1109\/CVPR42600.2020.00522"},{"key":"3171_CR8","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. Adv. Neural Inform. Process. Syst. 27 (2014)"},{"key":"3171_CR9","doi-asserted-by":"crossref","unstructured":"Guo, Z., Shao, M., Li, S.: Image-to-image translation using an offset-based multi-scale codes gan encoder. Visual Comput. 1\u201317 (2023)","DOI":"10.1007\/s00371-023-02810-4"},{"key":"3171_CR10","doi-asserted-by":"crossref","unstructured":"He, B., Gao, F., Ma, D., Shi, B., Duan, L.-Y.: Chipgan: a generative adversarial network for chinese ink wash painting style transfer. In: Proceedings of the 26th ACM international conference on Multimedia, pp. 1172\u20131180 (2018)","DOI":"10.1145\/3240508.3240655"},{"key":"3171_CR11","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local Nash equilibrium. Adv. Neural Inform. Process. Syst. 30 (2017)"},{"issue":"1","key":"3171_CR12","doi-asserted-by":"publisher","first-page":"1491","DOI":"10.1080\/17538947.2023.2202422","volume":"16","author":"S Hong","year":"2023","unstructured":"Hong, S., Shen, J., L\u00fc, G., Liu, X., Mao, Y., Sun, N., Tang, L.: Aesthetic style transferring method based on deep neural network between Chinese landscape painting and classical private garden\u2019s virtual scenario. Int. J. Digit. Earth 16(1), 1491\u20131509 (2023)","journal-title":"Int. J. Digit. Earth"},{"key":"3171_CR13","doi-asserted-by":"crossref","unstructured":"Howard, A., Sandler, M., Chu, G., Chen, L.-C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., Vasudevan, V., et\u00a0al.: Searching for mobilenetv3. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"3171_CR14","doi-asserted-by":"publisher","first-page":"1435","DOI":"10.1109\/TMM.2021.3065230","volume":"24","author":"J Huang","year":"2021","unstructured":"Huang, J., Liao, J., Kwong, S.: Unsupervised image-to-image translation via pre-trained stylegan2 network. IEEE Trans. Multim. 24, 1435\u20131448 (2021)","journal-title":"IEEE Trans. Multim."},{"key":"3171_CR15","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1501\u20131510 (2017)","DOI":"10.1109\/ICCV.2017.167"},{"key":"3171_CR16","doi-asserted-by":"crossref","unstructured":"Huang, X., Liu, M.-Y., Belongie, S., Kautz, J.: Multimodal unsupervised image-to-image translation. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 172\u2013189 (2018)","DOI":"10.1007\/978-3-030-01219-9_11"},{"key":"3171_CR17","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.-Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125\u20131134 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"3171_CR18","doi-asserted-by":"crossref","unstructured":"Lai, Y., Xu, H., Lin, C., Luo, T., Wang, L.: A two-stage and two-branch generative adversarial network-based underwater image enhancement. Visual Comput. 1\u201315 (2022)","DOI":"10.1007\/s00371-022-02580-5"},{"key":"3171_CR19","doi-asserted-by":"crossref","unstructured":"Lee, H.-Y., Tseng, H.-Y., Huang, J.-B., Singh, M., Yang, M.-H.: Diverse image-to-image translation via disentangled representations. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 35\u201351 (2018)","DOI":"10.1007\/978-3-030-01246-5_3"},{"issue":"11","key":"3171_CR20","doi-asserted-by":"publisher","first-page":"3577","DOI":"10.1007\/s00371-021-02188-1","volume":"38","author":"L Li","year":"2022","unstructured":"Li, L., Tang, J., Shao, Z., Tan, X., Ma, L.: Sketch-to-photo face generation based on semantic consistency preserving and similar connected component refinement. Vis. Comput. 38(11), 3577\u20133594 (2022)","journal-title":"Vis. Comput."},{"key":"3171_CR21","doi-asserted-by":"crossref","unstructured":"Li, X., Zhang, S., Hu, J., Cao, L., Hong, X., Mao, X., Huang, F., Wu, Y., Ji, R.: Image-to-image translation via hierarchical style disentanglement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8639\u20138648 (2021)","DOI":"10.1109\/CVPR46437.2021.00853"},{"key":"3171_CR22","unstructured":"Liu, M.-Y., Breuel, T., Kautz, J.: Unsupervised image-to-image translation networks. Adv. Neural Inform. Process. Syst. 30 (2017)"},{"key":"3171_CR23","doi-asserted-by":"crossref","unstructured":"Park, T., Liu, M.-Y., Wang, T.-C., Zhu, J.-Y.: Semantic image synthesis with spatially-adaptive normalization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2337\u20132346 (2019)","DOI":"10.1109\/CVPR.2019.00244"},{"issue":"3","key":"3171_CR24","doi-asserted-by":"publisher","first-page":"619","DOI":"10.1007\/s41095-022-0295-3","volume":"9","author":"Z Peng","year":"2023","unstructured":"Peng, Z., Wang, H., Weng, Y., Yang, Y., Shao, T.: Unsupervised image translation with distributional semantics awareness. Comput. Visual Media 9(3), 619\u2013631 (2023)","journal-title":"Comput. Visual Media"},{"issue":"2","key":"3171_CR25","doi-asserted-by":"publisher","first-page":"1700","DOI":"10.1109\/TPAMI.2022.3166687","volume":"45","author":"SR Richter","year":"2022","unstructured":"Richter, S.R., AlHaija, H.A., Vladlen, H.: Enhancing photorealism enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 45(2), 1700\u20131715 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3171_CR26","doi-asserted-by":"crossref","unstructured":"Saharia, C., Ho, J., Chan, W., Salimans, T., Fleet, D.J., Norouzi, M.: Image super-resolution via iterative refinement. IEEE Transactions on Pattern Analysis and Machine Intelligence (2022)","DOI":"10.1109\/TPAMI.2022.3204461"},{"key":"3171_CR27","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition (2014). arXiv:1409.1556"},{"key":"3171_CR28","doi-asserted-by":"crossref","unstructured":"Tan, W.R., Chan, C.S., Aguirre, H.E., Tanaka, K.: Artgan: Artwork synthesis with conditional categorical gans. In: 2017 IEEE International Conference on Image Processing (ICIP), pp. 3760\u20133764. IEEE (2017)","DOI":"10.1109\/ICIP.2017.8296985"},{"issue":"5","key":"3171_CR29","first-page":"6055","volume":"45","author":"H Tang","year":"2022","unstructured":"Tang, H., Torr, P.H.S., Sebe, N.: Multi-channel attention selection gans for guided image-to-image translation. IEEE Trans. Pattern Anal. Mach. Intell. 45(5), 6055\u20136071 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3171_CR30","doi-asserted-by":"publisher","first-page":"6664","DOI":"10.1109\/TIP.2022.3214336","volume":"31","author":"Z Bi","year":"2022","unstructured":"Bi, Z., Cao, B., Zuo, W., Hu, Q.: Learning a prototype discriminator with RBF for multimodal image synthesis. IEEE Trans. Image Process. 31, 6664\u20136678 (2022)","journal-title":"IEEE Trans. Image Process."},{"key":"3171_CR31","doi-asserted-by":"crossref","unstructured":"Cao, B., Bi, Z., Hu, Q., Zhang, H., Wang, N., Gao, X., Shen, D.: AutoEncoder-driven multimodal collaborative learning for medical image synthesis. Int. J. Comput. Vis. 1\u201320 (2023)","DOI":"10.1007\/s11263-023-01791-0"},{"key":"3171_CR32","doi-asserted-by":"crossref","unstructured":"Park, Taesung, Efros, Alexei A., Zhang, Richard, Zhu, Jun-Yan: Contrastive learning for unpaired image-to-image translation. In  Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part IX 16, pages 319\u2013345. Springer, (2020)","DOI":"10.1007\/978-3-030-58545-7_19"},{"key":"3171_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, J., Jiao, L., Ma, W., Liu, F., Liu, X., Li, L., Chen, P., Yang, S.: Transformer based conditional GAN for multimodal image fusion. IEEE Trans. Multim. (2023)","DOI":"10.1109\/TMM.2023.3243659"},{"key":"3171_CR34","first-page":"1","volume":"19","author":"Y Zhao","year":"2021","unstructured":"Zhao, Y., Shen, S., Hu, J., Li, Y., Pan, J.: Cloud removal using multimodal GAN with adversarial consistency loss. IEEE Geosci. Remote Sens. Lett. 19, 1\u20135 (2021)","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"3171_CR35","doi-asserted-by":"crossref","unstructured":"Tumanyan, N., Bar-Tal, O., Bagon, S., Dekel, T.: Splicing vit features for semantic appearance transfer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10748\u201310757 (2022)","DOI":"10.1109\/CVPR52688.2022.01048"},{"key":"3171_CR36","doi-asserted-by":"publisher","DOI":"10.1016\/j.aei.2023.101940","volume":"56","author":"S Varghese","year":"2023","unstructured":"Varghese, S., Hoskere, V.: Unpaired image-to-image translation of structural damage. Adv. Eng. Inform. 56, 101940 (2023)","journal-title":"Adv. Eng. Inform."},{"key":"3171_CR37","doi-asserted-by":"crossref","unstructured":"Wang, H., Li, P., Lang, X., Tao, D., Ma, J., Li, X.: Ftgan: A novel gan-based data augmentation method coupled time-frequency domain for imbalanced bearing fault diagnosis. IEEE Trans. Instrum. Measure. (2023)","DOI":"10.1109\/TIM.2023.3234095"},{"key":"3171_CR38","unstructured":"Wang, L., Sun, Y., Wang, Z.: Ccs-gan: a semi-supervised generative adversarial network for image classification. Visual Comput. 1\u201313 (2022)"},{"key":"3171_CR39","doi-asserted-by":"crossref","unstructured":"Wu, B., Dong, Q., Sun, W.: Automatic colorization of chinese ink painting combining multi-level features and generative adversarial networks. Fractals 2340144 (2023)","DOI":"10.1142\/S0218348X23401448"},{"key":"3171_CR40","doi-asserted-by":"crossref","unstructured":"Xie, Saining, Tu, Zhuowen: Holistically-nested edge detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1395-1403 (2015)","DOI":"10.1109\/ICCV.2015.164"},{"key":"3171_CR41","doi-asserted-by":"crossref","unstructured":"Xu, Shuai, Chang, Dongliang, Xie, Jiyang, Ma, Zhanyu: Grad-cam guided channel-spatial attention module for fine-grained visual classification. In  2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP), pages 1\u20136. IEEE, (2021)","DOI":"10.1109\/MLSP52302.2021.9596481"},{"key":"3171_CR42","doi-asserted-by":"crossref","unstructured":"Xue, Alice: End-to-end chinese landscape painting creation using generative adversarial networks. In  Proceedings of the IEEE\/CVF Winter conference on applications of computer vision, pages 3863\u20133871, (2021)","DOI":"10.1109\/WACV48630.2021.00391"},{"key":"3171_CR43","doi-asserted-by":"crossref","unstructured":"Zhang, Richard, Isola, Phillip, Efros, Alexei\u00a0A, Shechtman, Eli, Wang, Oliver: The unreasonable effectiveness of deep features as a perceptual metric. In  Proceedings of the IEEE conference on computer vision and pattern recognition, pages 586\u2013595, (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"issue":"3","key":"3171_CR44","doi-asserted-by":"publisher","first-page":"3144","DOI":"10.1109\/TII.2022.3160705","volume":"19","author":"X Zhang","year":"2022","unstructured":"Zhang, X., Fan, C., Xiao, Z., Zhao, L., Chen, H., Chang, X.: Random reconstructed unpaired image-to-image translation. IEEE Trans. Industr. Inf. 19(3), 3144\u20133154 (2022)","journal-title":"IEEE Trans. Industr. Inf."},{"issue":"4","key":"3171_CR45","first-page":"1283","volume":"39","author":"Y Zhang","year":"2023","unstructured":"Zhang, Y., Han, S., Zhang, Z., Wang, J., Bi, H.: Cf-gan: cross-domain feature fusion generative adversarial network for text-to-image synthesis. Vis. Comput. 39(4), 1283\u20131293 (2023)","journal-title":"Vis. Comput."},{"key":"3171_CR46","doi-asserted-by":"crossref","unstructured":"Zhou, Bolei, Khosla, Aditya, Lapedriza, Agata, Oliva, Aude, Torralba, Antonio: Learning deep features for discriminative localization. In  Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2921\u20132929, (2016)","DOI":"10.1109\/CVPR.2016.319"},{"key":"3171_CR47","doi-asserted-by":"crossref","unstructured":"Zhu, Jun-Yan, Park, Taesung, Isola, Phillip, Efros, Alexei\u00a0A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In  Proceedings of the IEEE international conference on computer vision, pages 2223\u20132232, (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"3171_CR48","doi-asserted-by":"crossref","unstructured":"Zhu M, He X, Wang N, Wang X, Gao X. All-to-key attention for arbitrary style transfer. In Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 23109\u201323119, (2023)","DOI":"10.1109\/ICCV51070.2023.02112"},{"key":"3171_CR49","doi-asserted-by":"publisher","first-page":"1820","DOI":"10.1007\/s11263-021-01442-2","volume":"129","author":"M Zhu","year":"2021","unstructured":"Zhu, M., Li, J., Wang, N., Gao, X.: Learning deep patch representation for probabilistic graphical model-based face sketch synthesis. Int. J. Comput. Vis. 129, 1820\u20131836 (2021)","journal-title":"Int. J. Comput. Vis."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03171-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-03171-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03171-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,13]],"date-time":"2024-08-13T15:18:11Z","timestamp":1723562291000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-03171-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,7]]},"references-count":49,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2024,9]]}},"alternative-id":["3171"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-03171-8","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2023,12,7]]},"assertion":[{"value":"5 November 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 December 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"No funding was received to assist with the preparation of this manuscript.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"Not applicable.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"The data that support the findings of this study are available on request from the corresponding author, Weibin Zhou, upon reasonable request.","order":6,"name":"Ethics","group":{"name":"EthicsHeading","label":"Code availability"}}]}}