{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T15:54:12Z","timestamp":1766159652903,"version":"3.40.3"},"reference-count":51,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2025,3,13]],"date-time":"2025-03-13T00:00:00Z","timestamp":1741824000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,3,13]],"date-time":"2025-03-13T00:00:00Z","timestamp":1741824000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62103432"],"award-info":[{"award-number":["62103432"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002858","name":"China Postdoctoral Science Foundation","doi-asserted-by":"publisher","award":["2022M721841"],"award-info":[{"award-number":["2022M721841"]}],"id":[{"id":"10.13039\/501100002858","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Young Talent Fund of the University Association for Science and Technology in Shannxi, China","award":["2021108"],"award-info":[{"award-number":["2021108"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2025,4]]},"DOI":"10.1007\/s40747-025-01827-7","type":"journal-article","created":{"date-parts":[[2025,3,13]],"date-time":"2025-03-13T08:10:42Z","timestamp":1741853442000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Discriminator guided visible-to-infrared image translation"],"prefix":"10.1007","volume":"11","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8545-6215","authenticated-orcid":false,"given":"Decao","family":"Ma","sequence":"first","affiliation":[]},{"given":"Juan","family":"Su","sequence":"additional","affiliation":[]},{"given":"Yong","family":"Xian","sequence":"additional","affiliation":[]},{"given":"Shaopeng","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,3,13]]},"reference":[{"issue":"8","key":"1827_CR1","doi-asserted-by":"crossref","first-page":"2939","DOI":"10.1007\/s00371-021-02166-7","volume":"38","author":"K Bayoudh","year":"2022","unstructured":"Bayoudh K, Knani R, Hamdaoui F, Mtibaa A (2022) A survey on deep multimodal learning for computer vision: advances, trends, applications, and datasets. Vis Comput 38(8):2939\u20132970","journal-title":"Vis Comput"},{"issue":"9","key":"1827_CR2","doi-asserted-by":"crossref","first-page":"17018","DOI":"10.3934\/mbe.2023759","volume":"20","author":"SF Abbasi","year":"2023","unstructured":"Abbasi SF, Abbasi QH, Saeed F, Alghamdi NS (2023) A convolutional neural network-based decision support system for neonatal quiet sleep detection. Math Biosci Eng 20(9):17018\u201317036","journal-title":"Math Biosci Eng"},{"key":"1827_CR3","doi-asserted-by":"publisher","first-page":"85929","DOI":"10.1109\/ACCESS.2024.3412412","volume":"12","author":"AA Siddique","year":"2024","unstructured":"Siddique AA, Raza A, Alshehri MS, Alasbali N, Abbasi SF (2024) Optimizing tumor classification through transfer learning and particle swarm optimization-driven feature extraction. IEEE Access 12:85929\u201385939. https:\/\/doi.org\/10.1109\/ACCESS.2024.3412412","journal-title":"IEEE Access"},{"key":"1827_CR4","doi-asserted-by":"publisher","first-page":"45762","DOI":"10.1109\/ACCESS.2024.3380816","volume":"12","author":"MS Alshehri","year":"2024","unstructured":"Alshehri MS, Saidani O, Alrayes FS, Abbasi SF, Ahmad J (2024) A self-attention-based deep convolutional neural networks for iiot networks intrusion detection. IEEE Access 12:45762\u201345772. https:\/\/doi.org\/10.1109\/ACCESS.2024.3380816","journal-title":"IEEE Access"},{"issue":"11","key":"1827_CR5","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2020) Generative adversarial networks. Commun ACM 63(11):139\u2013144","journal-title":"Commun ACM"},{"key":"1827_CR6","unstructured":"Arjovsky M, Chintala S, Bottou L (2017) Wasserstein generative adversarial networks. In: International conference on machine learning. PMLR, pp 214\u2013223"},{"key":"1827_CR7","doi-asserted-by":"crossref","unstructured":"Mao X, Li Q, Xie H, Lau RY, Wang Z, Paul\u00a0Smolley S (2017) Least squares generative adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2794\u20132802","DOI":"10.1109\/ICCV.2017.304"},{"key":"1827_CR8","volume":"110","author":"W Li","year":"2021","unstructured":"Li W, Fan L, Wang Z, Ma C, Cui X (2021) Tackling mode collapse in multi-generator GANs with orthogonal vectors. Pattern Recogn 110:107646","journal-title":"Pattern Recogn"},{"key":"1827_CR9","first-page":"37690","volume":"35","author":"Z Pan","year":"2022","unstructured":"Pan Z, Niu L, Zhang L (2022) UniGAN: reducing mode collapse in GANs using a uniform generator. Adv Neural Inf Process Syst 35:37690\u201337703","journal-title":"Adv Neural Inf Process Syst"},{"key":"1827_CR10","doi-asserted-by":"crossref","DOI":"10.1016\/j.artint.2023.103863","volume":"317","author":"M Allahyani","year":"2023","unstructured":"Allahyani M, Alsulami R, Alwafi T, Alafif T, Ammar H, Sabban S, Chen X (2023) DivGAN: a diversity enforcing generative adversarial network for mode collapse reduction. Artif Intell 317:103863","journal-title":"Artif Intell"},{"key":"1827_CR11","doi-asserted-by":"crossref","unstructured":"Hu X, Zhou X, Huang Q, Shi Z, Sun L, Li Q (2022) QS-ATTN: query-selected attention for contrastive learning in I2I translation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 18291\u201318300","DOI":"10.1109\/CVPR52688.2022.01775"},{"key":"1827_CR12","doi-asserted-by":"crossref","unstructured":"Suri S, Meshry M, Davis LS, Shrivastava A (2024) Grit: gan residuals for paired image-to-image translation. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision (WACV), pp 4965\u20134975","DOI":"10.1109\/WACV57701.2024.00489"},{"key":"1827_CR13","unstructured":"Mirza M, Osindero S (2014) Conditional generative adversarial nets. Preprint arXiv:1411.1784"},{"key":"1827_CR14","doi-asserted-by":"crossref","unstructured":"Kniaz VV, Knyaz VA, Hladuvka J, Kropatsch WG, Mizginov V (2018) Thermalgan: multimodal color-to-thermal image translation for person re-identification in multispectral dataset. In: Proceedings of the European conference on computer vision (ECCV) workshops","DOI":"10.1007\/978-3-030-11024-6_46"},{"key":"1827_CR15","doi-asserted-by":"publisher","unstructured":"Ma Y, Hua Y, Zuo Z (2021) Infrared image generation by pix2pix based on multi-receptive field feature fusion. In: 2021 international conference on control, automation and information sciences (ICCAIS), pp 1029\u20131036. https:\/\/doi.org\/10.1109\/ICCAIS52680.2021.9624500","DOI":"10.1109\/ICCAIS52680.2021.9624500"},{"key":"1827_CR16","doi-asserted-by":"crossref","first-page":"69","DOI":"10.1016\/j.patrec.2022.01.026","volume":"155","author":"MA \u00d6zkano\u011flu","year":"2022","unstructured":"\u00d6zkano\u011flu MA, Ozer S (2022) InfraGAN: a GAN architecture to transfer visible images to infrared domain. Pattern Recogn Lett 155:69\u201376","journal-title":"Pattern Recogn Lett"},{"key":"1827_CR17","first-page":"1","volume":"2023","author":"D Ma","year":"2023","unstructured":"Ma D, Xian Y, Li B, Li S, Zhang D (2023) Visible-to-infrared image translation based on an improved CGAN. Vis Comput 2023:1\u201310","journal-title":"Vis Comput"},{"key":"1827_CR18","doi-asserted-by":"publisher","first-page":"18199","DOI":"10.1109\/JSTARS.2024.3468456","volume":"17","author":"D Ma","year":"2024","unstructured":"Ma D, Li S, Su J, Xian Y, Zhang T (2024) Visible-to-infrared image translation for matching tasks. IEEE J Sel Top Appl Earth Observ Remote Sens 17:18199\u201318213. https:\/\/doi.org\/10.1109\/JSTARS.2024.3468456","journal-title":"IEEE J Sel Top Appl Earth Observ Remote Sens"},{"key":"1827_CR19","doi-asserted-by":"publisher","unstructured":"Lee D, Jeon M, Cho Y, Kim A (2023) Edge-guided multi-domain rgb-to-tir image translation for training vision tasks with challenging labels. In: 2023 IEEE international conference on robotics and automation (ICRA), pp 8291\u20138298. https:\/\/doi.org\/10.1109\/ICRA48891.2023.10161210","DOI":"10.1109\/ICRA48891.2023.10161210"},{"key":"1827_CR20","doi-asserted-by":"crossref","unstructured":"Isola P, Zhu J-Y, Zhou T, Efros AA (2017) Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1125\u20131134","DOI":"10.1109\/CVPR.2017.632"},{"key":"1827_CR21","doi-asserted-by":"crossref","unstructured":"Zhu J-Y, Park T, Isola P, Efros AA (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2223\u20132232","DOI":"10.1109\/ICCV.2017.244"},{"key":"1827_CR22","unstructured":"Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. Preprint arXiv:1503.02531"},{"key":"1827_CR23","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"1827_CR24","doi-asserted-by":"crossref","unstructured":"Cao H, Wang Y, Chen J, Jiang D, Zhang X, Tian Q, Wang M (2022) Swin-unet: Unet-like pure transformer for medical image segmentation. In: European conference on computer vision. Springer, pp 205\u2013218","DOI":"10.1007\/978-3-031-25066-8_9"},{"key":"1827_CR25","doi-asserted-by":"crossref","unstructured":"Zhang H, Zu K, Lu J, Zou Y, Meng D (2022) Epsanet: an efficient pyramid squeeze attention block on convolutional neural network. In: Proceedings of the Asian conference on computer vision, pp 1161\u20131177","DOI":"10.1007\/978-3-031-26313-2_33"},{"key":"1827_CR26","doi-asserted-by":"crossref","first-page":"702","DOI":"10.1007\/978-3-319-46487-9_43","volume-title":"Computer vision\u2014ECCV 2016","author":"C Li","year":"2016","unstructured":"Li C, Wand M (2016) Precomputed real-time texture synthesis with Markovian generative adversarial networks. In: Leibe B, Matas J, Sebe N, Welling M (eds) Computer vision\u2014ECCV 2016. Springer, Cham, pp 702\u2013716"},{"key":"1827_CR27","doi-asserted-by":"crossref","unstructured":"Wang T-C, Liu M-Y, Zhu J-Y, Tao A, Kautz J, Catanzaro B (2018) High-resolution image synthesis and semantic manipulation with conditional GANs. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8798\u20138807","DOI":"10.1109\/CVPR.2018.00917"},{"key":"1827_CR28","doi-asserted-by":"publisher","unstructured":"Sch\u00f6nfeld E, Schiele B, Khoreva A (2020) A u-net based discriminator for generative adversarial networks. In: 2020 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 8204\u20138213. https:\/\/doi.org\/10.1109\/CVPR42600.2020.00823","DOI":"10.1109\/CVPR42600.2020.00823"},{"key":"1827_CR29","doi-asserted-by":"crossref","unstructured":"Yun S, Park J, Lee K, Shin J (2020) Regularizing class-wise predictions via self-knowledge distillation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 13876\u201313885","DOI":"10.1109\/CVPR42600.2020.01389"},{"key":"1827_CR30","first-page":"1","volume":"31","author":"X Zhu","year":"2018","unstructured":"Zhu X, Gong S et al (2018) Knowledge distillation by on-the-fly native ensemble. Adv Neural Inform Process Syst 31:1","journal-title":"Adv Neural Inform Process Syst"},{"key":"1827_CR31","doi-asserted-by":"crossref","unstructured":"Nguyen-Meidine LT, Belal A, Kiran M, Dolz J, Blais-Morin L-A, Granger E (2021) Unsupervised multi-target domain adaptation through knowledge distillation. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 1339\u20131347","DOI":"10.1109\/WACV48630.2021.00138"},{"key":"1827_CR32","doi-asserted-by":"crossref","unstructured":"Gupta S, Hoffman J, Malik J (2016) Cross modal distillation for supervision transfer. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2827\u20132836","DOI":"10.1109\/CVPR.2016.309"},{"key":"1827_CR33","doi-asserted-by":"crossref","unstructured":"Chawla A, Yin H, Molchanov P, Alvarez J (2021) Data-free knowledge distillation for object detection. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 3289\u20133298","DOI":"10.1109\/WACV48630.2021.00333"},{"key":"1827_CR34","doi-asserted-by":"crossref","first-page":"10555","DOI":"10.1109\/TPAMI.2023.3257546","volume":"45","author":"Z Li","year":"2023","unstructured":"Li Z, Xu P, Chang X, Yang L, Zhang Y, Yao L, Chen X (2023) When object detection meets knowledge distillation: a survey. IEEE Trans Pattern Anal Mach Intell 45:10555\u201310579","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1827_CR35","doi-asserted-by":"crossref","unstructured":"Yuan J, Ge J, Wang Z, Liu Y (2023) Semi-supervised semantic segmentation with mutual knowledge distillation. In: Proceedings of the 31st ACM international conference on multimedia, pp 5436\u20135444","DOI":"10.1145\/3581783.3611906"},{"key":"1827_CR36","doi-asserted-by":"crossref","unstructured":"Yang C, Zhou H, An Z, Jiang X, Xu Y, Zhang Q (2022) Cross-image relational knowledge distillation for semantic segmentation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 12319\u201312328","DOI":"10.1109\/CVPR52688.2022.01200"},{"key":"1827_CR37","doi-asserted-by":"crossref","unstructured":"Cui K, Yu Y, Zhan F, Liao S, Lu S, Xing EP (2023) Kd-dlgan: data limited image generation via knowledge distillation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 3872\u20133882","DOI":"10.1109\/CVPR52729.2023.00377"},{"key":"1827_CR38","doi-asserted-by":"crossref","unstructured":"Liu Z, Mao H, Wu C-Y, Feichtenhofer C, Darrell T, Xie S (2022) A convnet for the 2020s. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 11976\u201311986","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"1827_CR39","doi-asserted-by":"crossref","unstructured":"Pinto F, Torr PH, Dokania P K (2022) An impartial take to the CNN vs. transformer robustness contest. In: European conference on computer vision. Springer, London, pp 466\u2013480","DOI":"10.1007\/978-3-031-19778-9_27"},{"key":"1827_CR40","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: a large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition. IEEE, pp 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"1827_CR41","first-page":"145","volume":"2020","author":"N Agarwal","year":"2021","unstructured":"Agarwal N, Sondhi A, Chopra K, Singh G (2021) Transfer learning: survey and classification. Smart Innov Commun Comput Sci Proc ICSICCS 2020:145\u2013155","journal-title":"Smart Innov Commun Comput Sci Proc ICSICCS"},{"key":"1827_CR42","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3149101","volume":"71","author":"W Xiao","year":"2022","unstructured":"Xiao W, Zhang Y, Wang H, Li F, Jin H (2022) Heterogeneous knowledge distillation for simultaneous infrared-visible image fusion and super-resolution. IEEE Trans Instrum Meas 71:1\u201315. https:\/\/doi.org\/10.1109\/TIM.2022.3149101","journal-title":"IEEE Trans Instrum Meas"},{"key":"1827_CR43","doi-asserted-by":"crossref","unstructured":"Hwang S, Park J, Kim N, Choi Y, So\u00a0Kweon I (2015) Multispectral pedestrian detection: benchmark dataset and baseline. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR)","DOI":"10.1109\/CVPR.2015.7298706"},{"key":"1827_CR44","doi-asserted-by":"crossref","unstructured":"Zhang H, Fromont E, Lefevre S, Avignon B (2020) Multispectral fusion for object detection with cyclic fuse-and-refine blocks. In: 2020 IEEE international conference on image processing (ICIP). IEEE, pp 276\u2013280","DOI":"10.1109\/ICIP40778.2020.9191080"},{"key":"1827_CR45","doi-asserted-by":"crossref","first-page":"0096","DOI":"10.34133\/remotesensing.0096","volume":"3","author":"Z Han","year":"2023","unstructured":"Han Z, Zhang Z, Zhang S, Zhang G, Mei S (2023) Aerial visible-to-infrared image translation: dataset, evaluation, and baseline. J Remote Sens 3:0096","journal-title":"J Remote Sens"},{"key":"1827_CR46","doi-asserted-by":"crossref","unstructured":"Hore A, Ziou D (2010) Image quality metrics: Psnr vs. ssim. In: 2010 20th international conference on pattern recognition. IEEE, pp 2366\u20132369","DOI":"10.1109\/ICPR.2010.579"},{"key":"1827_CR47","doi-asserted-by":"publisher","unstructured":"Wang Z, Simoncelli EP, Bovik AC (2003) Multiscale structural similarity for image quality assessment. In: The thrity-seventh asilomar conference on signals, systems and computers, 2003, vol 2, pp. 1398\u201314022. https:\/\/doi.org\/10.1109\/ACSSC.2003.1292216","DOI":"10.1109\/ACSSC.2003.1292216"},{"key":"1827_CR48","doi-asserted-by":"crossref","unstructured":"Zhang R, Isola P, Efros AA, Shechtman E, Wang O (2018) The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 86\u2013595","DOI":"10.1109\/CVPR.2018.00068"},{"key":"1827_CR49","first-page":"1","volume":"30","author":"M Heusel","year":"2017","unstructured":"Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S (2017) GANs trained by a two time-scale update rule converge to a local Nash equilibrium. Adv Neural Inform Process Syst 30:1","journal-title":"Adv Neural Inform Process Syst"},{"key":"1827_CR50","unstructured":"Bi\u0144kowski M, Sutherland DJ, Arbel M, Gretton A (2018) Demystifying mmd GANs. Preprint arXiv:1801.01401"},{"issue":"5","key":"1827_CR51","doi-asserted-by":"crossref","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li H, Wu X-J (2018) DenseFuse: a fusion approach to infrared and visible images. IEEE Trans Image Process 28(5):2614\u20132623","journal-title":"IEEE Trans Image Process"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01827-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-025-01827-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01827-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,30]],"date-time":"2025-03-30T21:25:58Z","timestamp":1743369958000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-025-01827-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,3,13]]},"references-count":51,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,4]]}},"alternative-id":["1827"],"URL":"https:\/\/doi.org\/10.1007\/s40747-025-01827-7","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"type":"print","value":"2199-4536"},{"type":"electronic","value":"2198-6053"}],"subject":[],"published":{"date-parts":[[2025,3,13]]},"assertion":[{"value":"4 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 February 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 March 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"208"}}