{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T17:09:20Z","timestamp":1776100160957,"version":"3.50.1"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2022,12,26]],"date-time":"2022-12-26T00:00:00Z","timestamp":1672012800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,12,26]],"date-time":"2022-12-26T00:00:00Z","timestamp":1672012800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62006240"],"award-info":[{"award-number":["62006240"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2023,8]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Unpaired image-to-image translation for the generation field has made much progress recently. However, these methods suffer from mode collapse because of the overfitting of the discriminator. To this end, we propose a straightforward method to construct a contrastive loss using the feature information of the discriminator output layer, which is named multi-feature contrastive learning (MCL). Our proposed method enhances the performance of the discriminator and solves the problem of model collapse by further leveraging contrastive learning. We perform extensive experiments on several open challenge datasets. Our method achieves state-of-the-art results compared with current methods. Finally, a series of ablation studies proved that our approach has better stability. In addition, our proposed method is also practical for single image translation tasks. Code is available at <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" ext-link-type=\"uri\" xlink:href=\"https:\/\/github.com\/gouayao\/MCL.\">https:\/\/github.com\/gouayao\/MCL.<\/jats:ext-link><\/jats:p>","DOI":"10.1007\/s40747-022-00924-1","type":"journal-article","created":{"date-parts":[[2022,12,26]],"date-time":"2022-12-26T14:06:42Z","timestamp":1672063602000},"page":"4111-4122","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":22,"title":["Multi-feature contrastive learning for unpaired image-to-image translation"],"prefix":"10.1007","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0105-3377","authenticated-orcid":false,"given":"Yao","family":"Gou","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3009-279X","authenticated-orcid":false,"given":"Min","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yu","family":"Song","sequence":"additional","affiliation":[]},{"given":"Yujie","family":"He","sequence":"additional","affiliation":[]},{"given":"Litao","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,12,26]]},"reference":[{"key":"924_CR1","doi-asserted-by":"crossref","unstructured":"Baek K, Choi Y, Uh Y, Yoo J, Shim H (2021) Rethinking the truly unsupervised image-to-image translation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 14154\u201314163","DOI":"10.1109\/ICCV48922.2021.01389"},{"key":"924_CR2","unstructured":"Benaim S, Wolf (2017) One-sided unsupervised domain mapping. In: NIPS, pp. 752\u2013762 http:\/\/papers.nips.cc\/paper\/6677-one-sided-unsupervised-domain-mapping"},{"key":"924_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2020.103014","volume":"74","author":"B Chaitanya","year":"2021","unstructured":"Chaitanya B, Mukherjee S (2021) Single image dehazing using improved cyclegan. J Vis Commun Image Represent 74:103014","journal-title":"J Vis Commun Image Represent"},{"key":"924_CR4","doi-asserted-by":"crossref","unstructured":"Chen Q, Koltun V (2017) Photographic image synthesis with cascaded refinement networks. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 1511\u20131520","DOI":"10.1109\/ICCV.2017.168"},{"key":"924_CR5","doi-asserted-by":"crossref","unstructured":"Chen J, Chen J, Chao H, Yang M (2018) Image blind denoising with generative adversarial network based noise modeling. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3155\u20133164","DOI":"10.1109\/CVPR.2018.00333"},{"key":"924_CR6","unstructured":"Chen T, Kornblith S, Norouzi M, Hinton GE (2020) A simple framework for contrastive learning of visual representations. CoRR. arXiv:2002.05709"},{"key":"924_CR7","doi-asserted-by":"crossref","unstructured":"Choi Y, Uh Y, Yoo J, Ha JW Stargan (2020) v2: Diverse image synthesis for multiple domains. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8188\u20138197","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"924_CR8","doi-asserted-by":"crossref","unstructured":"Cordts M, Omran M, Ramos S, Rehfeld T, Enzweiler M, Benenson R, Franke U, Roth S, Schiele B (2016) The cityscapes dataset for semantic urban scene understanding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3213\u20133223","DOI":"10.1109\/CVPR.2016.350"},{"key":"924_CR9","unstructured":"Dash A, Ye J, Wang G (2021) A review of generative adversarial networks (gans) and its applications in a wide variety of disciplines\u2014from medical to remote sensing"},{"key":"924_CR10","doi-asserted-by":"publisher","unstructured":"Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255 https:\/\/doi.org\/10.1109\/CVPR.2009.5206848","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"924_CR11","doi-asserted-by":"crossref","unstructured":"Fu H, Gong M, Wang C, Batmanghelich K, Zhang K, Tao D (2019) Geometry-consistent generative adversarial networks for one-sided unsupervised domain mapping. In: CVPR, pp. 2427\u20132436","DOI":"10.1109\/CVPR.2019.00253"},{"key":"924_CR12","doi-asserted-by":"crossref","unstructured":"Gatys LA, Ecker AS, Bethge M (2016) Image style transfer using convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2414\u20132423","DOI":"10.1109\/CVPR.2016.265"},{"key":"924_CR13","doi-asserted-by":"crossref","unstructured":"GM, H., Gourisaria, M.K., Pandey, M., Rautaray, S.S. (2020) A comprehensive survey and analysis of generative models in machine learning. Comput Sci Rev 38:100285","DOI":"10.1016\/j.cosrev.2020.100285"},{"key":"924_CR14","unstructured":"Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville AC, Bengio Y (2014) Generative adversarial nets. In: NIPS, pp. 2672\u20132680 . http:\/\/papers.nips.cc\/paper\/5423-generative-adversarial-nets"},{"key":"924_CR15","unstructured":"Gutmann M, Hyv\u00e4rinen A (2010) Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In: Y.W. Teh, M. Titterington (eds.) Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Proceedings of Machine Learning Research, vol. 9, pp. 297\u2013304. PMLR, Chia Laguna Resort, Sardinia, Italy https:\/\/proceedings.mlr.press\/v9\/gutmann10a.html"},{"key":"924_CR16","doi-asserted-by":"crossref","unstructured":"Han J, Shoeiby M, Petersson L, Armin MA (2021) Dual contrastive learning for unsupervised image-to-image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, pp. 746\u2013755","DOI":"10.1109\/CVPRW53098.2021.00084"},{"key":"924_CR17","doi-asserted-by":"crossref","unstructured":"He K, Fan H, Wu Y, Xie S, Girshick R (2020) Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9729\u20139738","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"924_CR18","unstructured":"Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S (2017) Gans trained by a two time-scale update rule converge to a local nash equilibrium. In: NIPS, pp. 6629\u20136640 http:\/\/papers.nips.cc\/paper\/7240-gans-trained-by-a-two-time-scale-update-rule-converge-to-a-local-nash-equilibrium"},{"key":"924_CR19","doi-asserted-by":"crossref","unstructured":"Huang X, Liu MY, Belongie S, Kautz J (2018) Multimodal unsupervised image-to-image translation. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 172\u2013189","DOI":"10.1007\/978-3-030-01219-9_11"},{"key":"924_CR20","doi-asserted-by":"crossref","unstructured":"Isola P, Zhu JY, Zhou T, Efros AA (2017) Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1125\u20131134","DOI":"10.1109\/CVPR.2017.632"},{"key":"924_CR21","unstructured":"Jeong J, Shin J (2021) Training gans with stronger augmentations via contrastive discriminator. In: International Conference on Learning Representations . https:\/\/openreview.net\/forum?id=eo6U4CAwVmg"},{"key":"924_CR22","doi-asserted-by":"publisher","first-page":"694","DOI":"10.1007\/978-3-319-46475-6_43","volume-title":"Computer Vision - ECCV 2016","author":"J Johnson","year":"2016","unstructured":"Johnson J, Alahi A, Fei-Fei L (2016) Perceptual losses for real-time style transfer and super-resolution. In: Leibe B, Matas J, Sebe N, Welling M (eds) Computer Vision - ECCV 2016. Springer International Publishing, Cham, pp 694\u2013711"},{"key":"924_CR23","doi-asserted-by":"crossref","unstructured":"Karras T, Laine S, Aittala M, Hellsten J, Lehtinen J, Aila T (2020) Analyzing and improving the image quality of stylegan. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8110\u20138119","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"924_CR24","unstructured":"Kim T, Cha M, Kim H, Lee JK, Kim J (2017) Learning to discover cross-domain relations with generative adversarial networks. In: D. Precup, Y.W. Teh (eds.) Proceedings of the 34th International Conference on Machine Learning, Proceedings of Machine Learning Research, vol. 70, pp. 1857\u20131865. PMLR https:\/\/proceedings.mlr.press\/v70\/kim17a.html"},{"key":"924_CR25","doi-asserted-by":"crossref","unstructured":"Kolkin N, Salavon J, Shakhnarovich G (2019) Style transfer by relaxed optimal transport and self-similarity. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10051\u201310060","DOI":"10.1109\/CVPR.2019.01029"},{"key":"924_CR26","doi-asserted-by":"crossref","unstructured":"Ledig C, Theis L, Huszar F, Caballero J, Cunningham A, Acosta A, Aitken A, Tejani A, Totz J, Wang Z, Shi W (2017) Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4681\u20134690","DOI":"10.1109\/CVPR.2017.19"},{"key":"924_CR27","doi-asserted-by":"crossref","unstructured":"Lee HY, Tseng HY, Huang JB, Singh M, Yang, MH (2018) Diverse image-to-image translation via disentangled representations. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 35\u201351","DOI":"10.1007\/978-3-030-01246-5_3"},{"key":"924_CR28","doi-asserted-by":"crossref","unstructured":"Li R, Pan J, Li Z, Tang J (2018) Single image dehazing via conditional generative adversarial network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8202\u20138211","DOI":"10.1109\/CVPR.2018.00856"},{"key":"924_CR29","doi-asserted-by":"publisher","unstructured":"Li T, Qian R, Dong C, Liu S, Yan Q, Zhu W, Lin L (2018) Beautygan: Instance-level facial makeup transfer with deep generative adversarial network. In: Proceedings of the 26th ACM International Conference on Multimedia, MM \u201918, p. 645-653. Association for Computing Machinery, New York, NY, USA https:\/\/doi.org\/10.1145\/3240508.3240618","DOI":"10.1145\/3240508.3240618"},{"key":"924_CR30","unstructured":"Liu MY, Breuel T, Kautz J (2017) Unsupervised image-to-image translation networks. In: Guyon I, Luxburg UV, . Bengio, H. Wallach, R. Fergus, S. Vishwanathan, R. Garnett (eds.) Advances in Neural Information Processing Systems, vol. 30. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2017\/file\/dc6a6489640ca02b0d42dabeb8e46bb7-Paper.pdf"},{"key":"924_CR31","doi-asserted-by":"crossref","unstructured":"Liu R, Ge Y, Choi CL, Wang X, Li H (2021). Divco: Diverse conditional image synthesis via contrastive generative adversarial network. CoRR. arXiv:2103.07893","DOI":"10.1109\/CVPR46437.2021.01611"},{"key":"924_CR32","doi-asserted-by":"crossref","unstructured":"Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3431\u20133440","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"924_CR33","doi-asserted-by":"publisher","unstructured":"Mao X, Li Q, Xie H, Lau RY, Wang Z, Smolley SP (2017) Least squares generative adversarial networks. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2813\u20132821. https:\/\/doi.org\/10.1109\/ICCV.2017.304","DOI":"10.1109\/ICCV.2017.304"},{"key":"924_CR34","doi-asserted-by":"crossref","unstructured":"Monday HN, Li J, Nneji GU, Nahar S, Hossin MA, Jackson J, Oluwasanmi A (2022) A wavelet convolutional capsule network with modified super resolution generative adversarial network for fault diagnosis and classification. Complex Intell Syst: 1\u201317","DOI":"10.1007\/s40747-022-00733-6"},{"key":"924_CR35","doi-asserted-by":"crossref","unstructured":"Park T, Liu MY, Wang TC, Zhu JY (2019) Semantic image synthesis with spatially-adaptive normalization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2337\u20132346","DOI":"10.1109\/CVPR.2019.00244"},{"key":"924_CR36","doi-asserted-by":"publisher","first-page":"319","DOI":"10.1007\/978-3-030-58545-7_19","volume-title":"Computer Vision - ECCV 2020","author":"T Park","year":"2020","unstructured":"Park T, Efros AA, Zhang R, Zhu JY (2020) Contrastive learning for unpaired image-to-image translation. In: Vedaldi A, Bischof H, Brox T, Frahm JM (eds) Computer Vision - ECCV 2020. Springer International Publishing, Cham, pp 319\u2013345"},{"key":"924_CR37","unstructured":"Salehi P, Chalechale A, Taghizadeh M (2020) Generative adversarial networks (gans): An overview of theoretical model, evaluation metrics, and recent developments. CoRR. arXiv:2005.13178"},{"key":"924_CR38","unstructured":"van den Oord A, Li Y, Vinyals O (2018) Representation learning with contrastive predictive coding. CoRR. arXiv:1807.03748"},{"key":"924_CR39","doi-asserted-by":"crossref","unstructured":"Wang TC, Liu MY, Zhu JY, Tao A, Kautz J, Catanzaro B (2018) High-resolution image synthesis and semantic manipulation with conditional gans. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8798\u20138807","DOI":"10.1109\/CVPR.2018.00917"},{"key":"924_CR40","doi-asserted-by":"crossref","unstructured":"Wang X, Yu K, Wu S, Gu J, Liu Y, Dong C, Qiao Y, Change Loy C (2018) Esrgan: Enhanced super-resolution generative adversarial networks. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops","DOI":"10.1007\/978-3-030-11021-5_5"},{"key":"924_CR41","doi-asserted-by":"crossref","unstructured":"Wang C, Zheng H, Yu Z, Zheng Z, Gu Z, Zheng B (2018) Discriminative region proposal adversarial networks for high-quality image-to-image translation. In: Proceedings of the European Conference on Computer Vision (ECCV)","DOI":"10.1007\/978-3-030-01246-5_47"},{"key":"924_CR42","doi-asserted-by":"crossref","unstructured":"Wu Z, Xiong Y, Yu SX, Lin D (2018) Unsupervised feature learning via non-parametric instance discrimination. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3733\u20133742","DOI":"10.1109\/CVPR.2018.00393"},{"key":"924_CR43","doi-asserted-by":"crossref","unstructured":"Yi Z, Zhang H, Tan P, Gong M (2017) Dualgan: Unsupervised dual learning for image-to-image translation. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 2849\u20132857","DOI":"10.1109\/ICCV.2017.310"},{"key":"924_CR44","doi-asserted-by":"crossref","unstructured":"Yoo J, Uh Y, Chun S, Kang B, Ha JW (2019) Photorealistic style transfer via wavelet transforms. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 9036\u20139045","DOI":"10.1109\/ICCV.2019.00913"},{"key":"924_CR45","doi-asserted-by":"crossref","unstructured":"Yu F, Koltun V, Funkhouser T (2017) Dilated residual networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 472\u2013480","DOI":"10.1109\/CVPR.2017.75"},{"key":"924_CR46","doi-asserted-by":"publisher","first-page":"649","DOI":"10.1007\/978-3-319-46487-9_40","volume-title":"Computer Vision - ECCV 2016","author":"R Zhang","year":"2016","unstructured":"Zhang R, Isola P, Efros AA (2016) Colorful image colorization. In: Leibe B, Matas J, Sebe N, Welling M (eds) Computer Vision - ECCV 2016. Springer International Publishing, Cham, pp 649\u2013666"},{"key":"924_CR47","doi-asserted-by":"crossref","unstructured":"Zhu JY, Park T, Isola P, Efros AA (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 2223\u20132232","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-022-00924-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-022-00924-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-022-00924-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,27]],"date-time":"2023-07-27T13:21:17Z","timestamp":1690464077000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-022-00924-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,12,26]]},"references-count":47,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2023,8]]}},"alternative-id":["924"],"URL":"https:\/\/doi.org\/10.1007\/s40747-022-00924-1","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,12,26]]},"assertion":[{"value":"20 January 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 November 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 December 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}