{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T19:40:09Z","timestamp":1768074009179,"version":"3.49.0"},"reference-count":62,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T00:00:00Z","timestamp":1751328000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T00:00:00Z","timestamp":1751328000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100013091","name":"Science and Technology Major Project of Guangxi","doi-asserted-by":"publisher","award":["GUIKE AA24206025"],"award-info":[{"award-number":["GUIKE AA24206025"]}],"id":[{"id":"10.13039\/501100013091","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100013091","name":"Science and Technology Major Project of Guangxi","doi-asserted-by":"publisher","award":["GUIKE AA23062038"],"award-info":[{"award-number":["GUIKE AA23062038"]}],"id":[{"id":"10.13039\/501100013091","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100011785","name":"Science and Technology Department of Guangxi Zhuang Autonomous Region","doi-asserted-by":"publisher","award":["GUIKE AD22080061"],"award-info":[{"award-number":["GUIKE AD22080061"]}],"id":[{"id":"10.13039\/501100011785","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100017691","name":"Guangxi Key Research and Development Program","doi-asserted-by":"publisher","award":["GUIKE AB23026120"],"award-info":[{"award-number":["GUIKE AB23026120"]}],"id":[{"id":"10.13039\/501100017691","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100017691","name":"Guangxi Key Research and Development Program","doi-asserted-by":"publisher","award":["GUIKE ZY22096026"],"award-info":[{"award-number":["GUIKE ZY22096026"]}],"id":[{"id":"10.13039\/501100017691","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62061010"],"award-info":[{"award-number":["62061010"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62161007"],"award-info":[{"award-number":["62161007"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100009329","name":"Scientific Research and Technology Development Program of Guangxi Zhuang Autonomous Region","doi-asserted-by":"publisher","award":["20210222-1"],"award-info":[{"award-number":["20210222-1"]}],"id":[{"id":"10.13039\/501100009329","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Innovation Project of Guangxi Graduate Education","award":["YCSW2024346"],"award-info":[{"award-number":["YCSW2024346"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s40747-025-01986-7","type":"journal-article","created":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T03:46:54Z","timestamp":1751341614000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Infrared and visible image fusion network based on multistage progressive injection"],"prefix":"10.1007","volume":"11","author":[{"given":"Kaixuan","family":"Chang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0217-9115","authenticated-orcid":false,"given":"Jianhua","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Xiyan","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Shitao","family":"Bao","sequence":"additional","affiliation":[]},{"given":"Huansheng","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,1]]},"reference":[{"key":"1986_CR1","doi-asserted-by":"publisher","unstructured":"Wu YF, Yang R, Lu QS et al (2024) Infrared and visible light image fusion: statistical analysis, deep learning methods, and future perspectives. Prog Laser Optoelectron 1\u201326. https:\/\/doi.org\/10.3788\/LOP232360","DOI":"10.3788\/LOP232360"},{"key":"1986_CR2","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2021.102488","volume":"66","author":"J Fu","year":"2021","unstructured":"Fu J, Li W, Du J et al (2021) A multiscale residual pyramid attention network for medical image fusion. Biomed Signal Process Control 66:102488. https:\/\/doi.org\/10.1016\/j.bspc.2021.102488","journal-title":"Biomed Signal Process Control"},{"key":"1986_CR3","doi-asserted-by":"publisher","unstructured":"Xu H, Ma J, Yuan J et al (2022) RFNet: unsupervised network for mutually reinforcing multi-modal image registration and fusion. In: 2022 IEEE\/CVF Conf Comput Vis Pattern Recognit(CVPR), pp 19647\u201319656. https:\/\/doi.org\/10.1016\/j.bspc.2021.102488","DOI":"10.1016\/j.bspc.2021.102488"},{"issue":"8","key":"1986_CR4","doi-asserted-by":"publisher","first-page":"5026","DOI":"10.1109\/TCSVT.2022.3144455","volume":"32","author":"J Liu","year":"2022","unstructured":"Liu J, Shang J, Liu R et al (2022) Attention-guided global-local adversarial learning for detail-preserving multi-exposure image fusion. IEEE Trans Circ Syst Video Technol 32(8):5026\u20135040. https:\/\/doi.org\/10.1109\/TCSVT.2022.3144455","journal-title":"IEEE Trans Circ Syst Video Technol"},{"key":"1986_CR5","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"H Zhang","year":"2021","unstructured":"Zhang H, Le Z, Shao Z et al (2021) MFF-GAN: an unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf Fusion 66:40\u201353. https:\/\/doi.org\/10.1016\/j.inffus.2020.08.022","journal-title":"Inf Fusion"},{"key":"1986_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2021.3090585","volume":"60","author":"J Wang","year":"2022","unstructured":"Wang J, Shao Z, Huang X et al (2022) A dual-path fusion network for pan-sharpening. IEEE Trans Geosci Remote Sens 60:1\u201314. https:\/\/doi.org\/10.1109\/TGRS.2021.3090585","journal-title":"IEEE Trans Geosci Remote Sens"},{"key":"1986_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2022.104435","volume":"127","author":"H Tang","year":"2022","unstructured":"Tang H, Liu G, Tang L et al (2022) MdedFusion: a multi-level detail enhancement decomposition method for infrared and visible image fusion. Infrared Phys Technol 127:104435. https:\/\/doi.org\/10.1016\/j.infrared.2022.104435","journal-title":"Infrared Phys Technol"},{"issue":"12","key":"1986_CR8","doi-asserted-by":"publisher","first-page":"2162","DOI":"10.3390\/electronics9122162","volume":"9","author":"C Sun","year":"2020","unstructured":"Sun C, Zhang C, Xiong N (2020) Infrared and visible image fusion techniques based on deep learning: a review. Electronics 9(12):2162. https:\/\/doi.org\/10.3390\/electronics9122162","journal-title":"Electronics"},{"issue":"8","key":"1986_CR9","doi-asserted-by":"publisher","first-page":"10535","DOI":"10.1109\/TPAMI.2023.3261282","volume":"45","author":"X Zhang","year":"2023","unstructured":"Zhang X, Demiris Y (2023) Visible and infrared image fusion using deep learning. IEEE Trans Pattern Anal Mach Intell 45(8):10535\u201310554. https:\/\/doi.org\/10.1109\/TPAMI.2023.3261282","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1986_CR10","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"J Ma","year":"2019","unstructured":"Ma J, Ma Y, Li C (2019) Infrared and visible image fusion methods and applications: a survey. Inf Fusion 45:153\u2013178. https:\/\/doi.org\/10.1016\/j.inffus.2018.02.004","journal-title":"Inf Fusion"},{"key":"1986_CR11","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen J, Li X, Luo L et al (2020) Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf Sci 508:64\u201378. https:\/\/doi.org\/10.1016\/j.ins.2019.08.066","journal-title":"Inf Sci"},{"key":"1986_CR12","doi-asserted-by":"publisher","DOI":"10.3788\/gzxb20225109.0910002","author":"M Wang","year":"2022","unstructured":"Wang M, Wang X, Zhang C (2022) Infrared and visible image fusion algorithm based on dynamic range compression enhancement and NSST. Acta Photonica Sinica. https:\/\/doi.org\/10.3788\/gzxb20225109.0910002","journal-title":"Acta Photonica Sinica"},{"issue":"7","key":"1986_CR13","first-page":"705","volume":"45","author":"Z Long","year":"2023","unstructured":"Long Z, Deng Y, Wang R et al (2023) Infrared and visible image fusion based on saliency detection and latent low-rank representation. Infrared Technol 45(7):705\u2013713","journal-title":"Infrared Technol"},{"key":"1986_CR14","doi-asserted-by":"publisher","first-page":"101","DOI":"10.3969\/j.issn.1671-637X.2021.11.021","volume":"28","author":"X Du","year":"2021","unstructured":"Du X, Sun XJ, Li CA (2021) Multi-space transformed underwater image enhancement method based on weighted fusion. Electro-Optic Control 28:101\u2013105. https:\/\/doi.org\/10.3969\/j.issn.1671-637X.2021.11.021","journal-title":"Electro-Optic Control"},{"key":"1986_CR15","unstructured":"Yuan TH, Gan ZL (2024) Infrared visible light depth expanded image fusion network based on jointly enhanced image pairs. Comput Sci 1\u201315"},{"key":"1986_CR16","doi-asserted-by":"publisher","first-page":"3","DOI":"10.11834\/jig.220422","volume":"28","author":"LF Tang","year":"2023","unstructured":"Tang LF, Zhang H, Xu H (2023) Review of image fusion methods based on deep learning. Chin J Image Gr 28:3\u201336","journal-title":"Chin J Image Gr"},{"key":"1986_CR17","doi-asserted-by":"publisher","first-page":"64","DOI":"10.3778\/j.issn.1002-8331.2203-0600","volume":"58","author":"XD An","year":"2022","unstructured":"An XD, Li YL, Wang F (2022) Review of infrared and visible light fusion algorithms for vehicle driver assistance system. Compu Eng Appl 58:64\u201375. https:\/\/doi.org\/10.3778\/j.issn.1002-8331.2203-0600","journal-title":"Compu Eng Appl"},{"issue":"C","key":"1986_CR18","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1016\/j.inffus.2015.11.003","volume":"30","author":"Z Zhou","year":"2016","unstructured":"Zhou Z, Wang B, Li S et al (2016) Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with gaussian and bilateral filters. Inf Fusion 30(C):15\u201326. https:\/\/doi.org\/10.1016\/j.inffus.2015.11.003","journal-title":"Inf Fusion"},{"key":"1986_CR19","doi-asserted-by":"publisher","unstructured":"Liu J, Fan X, Huang Z et al (2022) Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: 2022 IEEE\/CVF Conf Comput Vision Pattern Recognit (CVPR), pp 5792\u20135801. https:\/\/doi.org\/10.1109\/CVPR52688.2022.00571","DOI":"10.1109\/CVPR52688.2022.00571"},{"issue":"12","key":"1986_CR20","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","volume":"69","author":"H Li","year":"2020","unstructured":"Li H, Wu X, Durrani T (2020) NestFuse: an infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models. IEEE Trans Instrum Meas 69(12):9645\u20139656. https:\/\/doi.org\/10.1109\/TIM.2020.3005230","journal-title":"IEEE Trans Instrum Meas"},{"key":"1986_CR21","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang L, Yuan J, Ma J (2022) Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inf Fusion 82:28\u201342. https:\/\/doi.org\/10.1016\/j.inffus.2021.12.004","journal-title":"Inf Fusion"},{"issue":"9","key":"1986_CR22","first-page":"840","volume":"43","author":"J Yang","year":"2021","unstructured":"Yang J, Liu W, Cheng Y (2021) Asymmetric infrared and visible image fusion based on contrast pyramid and bilateral filtering. Infrared Technol 43(9):840\u2013844","journal-title":"Infrared Technol"},{"key":"1986_CR23","unstructured":"Li JX (2023) Small object detection algorithm based on deep learning. Master\u2019s thesis, Xi\u2019an Petroleum University"},{"key":"1986_CR24","unstructured":"Niu ZH (2023) Research on the infrared and visible light image fusion algorithm based on multiscale transform. Master\u2019s thesis, Qingdao University of Technology"},{"key":"1986_CR25","unstructured":"Yang C (2023) Research on infrared micro-light image fusion technology based on multi-scale segmentation and detail enhancement. Master\u2019s thesis, Changchun University of Science and Technology"},{"key":"1986_CR26","doi-asserted-by":"publisher","first-page":"11066","DOI":"10.1109\/ACCESS.2024.3354169","volume":"12","author":"Z Wang","year":"2024","unstructured":"Wang Z, Zhou D, Li Z et al (2024) Underwater image enhancement via adaptive color correction and stationary wavelet detail enhancement. IEEE Access 12:11066\u201311082. https:\/\/doi.org\/10.1109\/ACCESS.2024.3354169","journal-title":"IEEE Access"},{"key":"1986_CR27","unstructured":"Liu G (2023) Image saliency object detection study based on multiple feature fusion. Master\u2019s thesis, Minnan Normal University"},{"issue":"11","key":"1986_CR28","first-page":"1072","volume":"42","author":"L Wu","year":"2020","unstructured":"Wu L, Chen N, Liao X (2020) Infrared image enhancement with region-adaptive multi-scale intense light fusion. Infrared Technol 42(11):1072\u20131076+1080","journal-title":"Infrared Technol"},{"key":"1986_CR29","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2021.3056645","volume":"70","author":"H Xu","year":"2021","unstructured":"Xu H, Wang X, Ma J (2021) DRF: disentangled representation for visible and infrared image fusion. IEEE Trans Instrum Meas 70:1\u201313. https:\/\/doi.org\/10.1109\/TIM.2021.3056645","journal-title":"IEEE Trans Instrum Meas"},{"issue":"5","key":"1986_CR30","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2019","unstructured":"Li H, Wu X (2019) DenseFuse: a fusion approach to infrared and visible images. IEEE Trans Image Process 28(5):2614\u20132623. https:\/\/doi.org\/10.1109\/TIP.2018.2887342","journal-title":"IEEE Trans Image Process"},{"key":"1986_CR31","doi-asserted-by":"publisher","unstructured":"Zhao Z, Xu S, Zhang C et al (2020) DIDFuse: deep image decomposition for infrared and visible image fusion. In: Bessiere C (ed) Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20. In: International Joint Conferences on Artificial Intelligence Organization pp 970\u2013976, https:\/\/doi.org\/10.24963\/ijcai.2020\/135","DOI":"10.24963\/ijcai.2020\/135"},{"key":"1986_CR32","doi-asserted-by":"publisher","first-page":"824","DOI":"10.1109\/TCI.2021.3100986","volume":"7","author":"H Xu","year":"2021","unstructured":"Xu H, Zhang H, Ma J (2021) Classification saliency-based rule for visible and infrared image fusion. IEEE Trans Comput Imaging 7:824\u2013836. https:\/\/doi.org\/10.1109\/TCI.2021.3100986","journal-title":"IEEE Trans Comput Imaging"},{"key":"1986_CR33","doi-asserted-by":"publisher","first-page":"4070","DOI":"10.1109\/TIP.2021.3069339","volume":"30","author":"H Li","year":"2021","unstructured":"Li H, Cen Y, Liu Y et al (2021) Different input resolutions and arbitrary output resolution: a meta learning-based deep framework for infrared and visible image fusion. IEEE Trans Image Process 30:4070\u20134083. https:\/\/doi.org\/10.1109\/TIP.2021.3069339","journal-title":"IEEE Trans Image Process"},{"key":"1986_CR34","doi-asserted-by":"crossref","unstructured":"Wang D, Liu J, Fan X et al (2022) Unsupervised misaligned infrared and visible image fusion via cross-modality image generation and registration. arXiv:2205.11876","DOI":"10.24963\/ijcai.2022\/487"},{"key":"1986_CR35","doi-asserted-by":"publisher","unstructured":"Sun Y, Cao B, Zhu P et al (2022) DetFusion: a detection-driven infrared and visible image fusion network. In: Proceedings of the 30th ACM International Conference of Multimedia https:\/\/doi.org\/10.1145\/3503161.3547902","DOI":"10.1145\/3503161.3547902"},{"key":"1986_CR36","doi-asserted-by":"publisher","first-page":"477","DOI":"10.1016\/j.inffus.2022.10.034","volume":"91","author":"Linfeng Tang","year":"2023","unstructured":"Tang Linfeng, Xiang Xinyu, Zhang Hao, Gong Meiqi, Ma Jiayi (2023) DIVFusion: Darkness-free infrared and visible image fusion. Inform Fusion 91:477\u2013493. https:\/\/doi.org\/10.1016\/j.inffus.2022.10.034","journal-title":"Inform Fusion"},{"key":"1986_CR37","doi-asserted-by":"publisher","unstructured":"Xu H, Liang P, Yu W et al (2019) Learning a generative model for fusing infrared and visible images via conditional generative adversarial network with dual discriminators. Int Joint Conf Artificial Intell https:\/\/doi.org\/10.24963\/ijcai.2019\/549","DOI":"10.24963\/ijcai.2019\/549"},{"key":"1986_CR38","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"Jiayi Ma","year":"2019","unstructured":"Ma Jiayi, Yu Wei, Liang Pengwei, Li Chang, Jiang Junjun (2019) FusionGAN: A generative adversarial network for infrared and visible image fusion. Inform Fusion 48:11\u201326. https:\/\/doi.org\/10.1016\/j.inffus.2018.09.004","journal-title":"Inform Fusion"},{"key":"1986_CR39","doi-asserted-by":"publisher","first-page":"635","DOI":"10.1109\/TMM.2021.3129609","volume":"25","author":"Huabing Zhou","year":"2023","unstructured":"Zhou Huabing, Wu Wei, Zhang Yanduo, Ma Jiayi, Ling Haibin (2023) Semantic-supervised infrared and visible image fusion via a dual-discriminator generative adversarial network. IEEE Trans Multimed 25:635\u2013648. https:\/\/doi.org\/10.1109\/TMM.2021.3129609","journal-title":"IEEE Trans Multimed"},{"key":"1986_CR40","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2020.3038013","volume":"70","author":"J Ma","year":"2021","unstructured":"Ma J, Zhang H, Shao Z et al (2021) GANMcC: a generative adversarial network with multiclassification constraints for infrared and visible image fusion. IEEE Trans Instrum Meas 70:1\u201314. https:\/\/doi.org\/10.1109\/TIM.2020.3038013","journal-title":"IEEE Trans Instrum Meas"},{"key":"1986_CR41","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101870","volume":"99","author":"L Tang","year":"2023","unstructured":"Tang L, Zhang H, Xu H et al (2023) Rethinking the necessity of image fusion in high-level vision tasks: a practical infrared and visible image fusion network based on progressive semantic injection and scene fidelity. Inf Fusion 99:101870. https:\/\/doi.org\/10.1016\/j.inffus.2023.101870","journal-title":"Inf Fusion"},{"issue":"1","key":"1986_CR42","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2022","unstructured":"Xu H, Ma J, Jiang J et al (2022) U2Fusion: a unified unsupervised image fusion network. IEEE Trans Pattern Anal Mach Intell 44(1):502\u2013518. https:\/\/doi.org\/10.1109\/TPAMI.2020.3012548","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1986_CR43","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1016\/j.dib.2017.09.038","volume":"15","author":"A Toet","year":"2017","unstructured":"Toet A (2017) The tno multiband image data collection. Data Brief 15:249\u2013251. https:\/\/doi.org\/10.1016\/j.dib.2017.09.038","journal-title":"Data Brief"},{"key":"1986_CR44","doi-asserted-by":"publisher","unstructured":"Ha Q, Watanabe K, Karasawa T et al (2017) MFNet: towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In: 2017 IEEE\/RSJ Int Conf Intell Robots Syst (IROS), pp 5108\u20135115. https:\/\/doi.org\/10.1109\/IROS.2017.8206396","DOI":"10.1109\/IROS.2017.8206396"},{"issue":"23","key":"1986_CR45","doi-asserted-by":"publisher","first-page":"6480","DOI":"10.1364\/AO.55.006480","volume":"55","author":"Z Zhou","year":"2016","unstructured":"Zhou Z, Dong M, Xie X et al (2016) Fusion of infrared and visible images for night-vision context enhancement. Appl Opt 55(23):6480\u201390. https:\/\/doi.org\/10.1364\/AO.55.006480","journal-title":"Appl Opt"},{"issue":"12","key":"1986_CR46","doi-asserted-by":"publisher","first-page":"3064","DOI":"10.1016\/j.infrared.2021.103938","volume":"58","author":"W Tan","year":"2019","unstructured":"Tan W, Zhou H, Song J et al (2019) Infrared and visible image perceptive fusion through multi-level Gaussian curvature filtering image decomposition. Appl Opt 58(12):3064\u20133073. https:\/\/doi.org\/10.1016\/j.infrared.2021.103938","journal-title":"Appl Opt"},{"key":"1986_CR47","doi-asserted-by":"publisher","unstructured":"Liu G, Yan S (2011) Latent low-rank representation for subspace segmentation and feature extraction. In: 2011 Int Conf Comput Vision pp 1615\u20131622. https:\/\/doi.org\/10.1109\/ICCV.2011.6126422","DOI":"10.1109\/ICCV.2011.6126422"},{"issue":"1","key":"1986_CR48","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1109\/TCSVT.2021.3056725","volume":"32","author":"J Liu","year":"2022","unstructured":"Liu J, Fan X, Jiang J et al (2022) Learning a deep multi-scale feature ensemble and an edge-attention guidance for image fusion. IEEE Trans Circ Syst Video Technol 32(1):105\u2013119. https:\/\/doi.org\/10.1109\/TCSVT.2021.3056725","journal-title":"IEEE Trans Circ Syst Video Technol"},{"key":"1986_CR49","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","volume":"83\u201384","author":"L Tang","year":"2022","unstructured":"Tang L, Yuan J, Zhang H et al (2022) PIAFusion: a progressive infrared and visible image fusion network based on illumination aware. Inf Fusion 83\u201384:79\u201392. https:\/\/doi.org\/10.1016\/j.inffus.2022.03.007","journal-title":"Inf Fusion"},{"key":"1986_CR50","doi-asserted-by":"publisher","first-page":"1134","DOI":"10.1109\/TCI.2021.3119954","volume":"7","author":"H Zhang","year":"2021","unstructured":"Zhang H, Yuan J, Tian X et al (2021) GAN-FM: infrared and visible image fusion using gan with full-scale skip connection and dual Markovian discriminators. IEEE Trans Comput Imaging 7:1134\u20131147. https:\/\/doi.org\/10.1109\/TCI.2021.3119954","journal-title":"IEEE Trans Comput Imaging"},{"issue":"7","key":"1986_CR51","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma J, Tang L, Fan F et al (2022) SwinFusion: cross-domain long-range learning for general image fusion via Swin transformer. IEEE\/CAA J Autom Sinica 9(7):1200\u20131217. https:\/\/doi.org\/10.1109\/JAS.2022.105686","journal-title":"IEEE\/CAA J Autom Sinica"},{"issue":"7","key":"1986_CR52","doi-asserted-by":"publisher","first-page":"3159","DOI":"10.1109\/TCSVT.2023.3234340","volume":"33","author":"W Tang","year":"2023","unstructured":"Tang W, He F, Liu Y et al (2023) DATFuse: infrared and visible image fusion via dual attention transformer. IEEE Trans Circ Syst Video Technol 33(7):3159\u20133172. https:\/\/doi.org\/10.1109\/TCSVT.2023.3234340","journal-title":"IEEE Trans Circ Syst Video Technol"},{"key":"1986_CR53","doi-asserted-by":"publisher","first-page":"1350","DOI":"10.1016\/j.infrared.2023.104701","volume":"131","author":"X Li","year":"2023","unstructured":"Li X, Tan H, Zhou F et al (2023) Infrared and visible image fusion based on domain transform filtering and sparse representation. Infrared Phys Technol 131:1350\u20134495. https:\/\/doi.org\/10.1016\/j.infrared.2023.104701","journal-title":"Infrared Phys Technol"},{"key":"1986_CR54","first-page":"1","volume":"73","author":"J Chen","year":"2024","unstructured":"Chen J, Yang L, Liu W et al (2024) LENFusion: a joint low-light enhancement and fusion network for nighttime infrared and visible image fusion. IEEE Trans Instrum Meas 73:1\u201315","journal-title":"IEEE Trans Instrum Meas"},{"key":"1986_CR55","doi-asserted-by":"publisher","unstructured":"Zhang X, Ye P, Xiao G (2020) VIFB: a visible and infrared image fusion benchmark. In: 2020 IEEE\/CVF Conf Comput Vision Pattern Recognit Workshops (CVPRW) pp 468\u2013478. https:\/\/doi.org\/10.1109\/CVPRW50498.2020.00060","DOI":"10.1109\/CVPRW50498.2020.00060"},{"key":"1986_CR56","doi-asserted-by":"publisher","DOI":"10.1117\/1.2945910","volume":"2","author":"J Roberts","year":"2008","unstructured":"Roberts J, Aardt J, Ahmed F (2008) Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J Appl Remote Sens 2:023522. https:\/\/doi.org\/10.1117\/1.2945910","journal-title":"J Appl Remote Sens"},{"issue":"12","key":"1986_CR57","doi-asserted-by":"publisher","first-page":"2959","DOI":"10.1109\/26.477498","volume":"43","author":"A Eskicioglu","year":"1995","unstructured":"Eskicioglu A, Fisher P (1995) Image quality measures and their performance. IEEE Trans Commun 43(12):2959\u20132965. https:\/\/doi.org\/10.1109\/26.477498","journal-title":"IEEE Trans Commun"},{"key":"1986_CR58","doi-asserted-by":"publisher","first-page":"199","DOI":"10.1016\/J.OPTCOM.2014.12.032","volume":"341","author":"G Cui","year":"2015","unstructured":"Cui G, Feng H, Xu Z et al (2015) Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Opt Commun 341:199\u2013209. https:\/\/doi.org\/10.1016\/J.OPTCOM.2014.12.032","journal-title":"Opt Commun"},{"key":"1986_CR59","doi-asserted-by":"publisher","first-page":"355","DOI":"10.1088\/0957-0233\/8\/4\/002","volume":"8","author":"Y Rao","year":"1997","unstructured":"Rao Y (1997) In-fibre bragg grating sensors. Meas Sci Technol 8:355\u2013375. https:\/\/doi.org\/10.1088\/0957-0233\/8\/4\/002","journal-title":"Meas Sci Technol"},{"issue":"2","key":"1986_CR60","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1016\/j.inffus.2011.08.002","volume":"14","author":"Y Han","year":"2013","unstructured":"Han Y, Cai Y, Cao Y et al (2013) A new image fusion performance metric based on visual information fidelity. Inf Fusion 14(2):127\u2013135. https:\/\/doi.org\/10.1016\/j.inffus.2011.08.002","journal-title":"Inf Fusion"},{"key":"1986_CR61","doi-asserted-by":"publisher","DOI":"10.1117\/1.2009764","volume":"44","author":"V Petrovic","year":"2005","unstructured":"Petrovic V, Xydeas C (2005) Objective evaluation of signal-level image fusion performance. Opt Eng 44:087003. https:\/\/doi.org\/10.1117\/1.2009764","journal-title":"Opt Eng"},{"issue":"10","key":"1986_CR62","doi-asserted-by":"publisher","first-page":"6700","DOI":"10.1109\/TCSVT.2022.3168279","volume":"32","author":"Y Sun","year":"2022","unstructured":"Sun Y, Cao B, Zhu P et al (2022) Drone-based rgb-infrared cross-modality vehicle detection via uncertainty-aware learning. IEEE Trans Circ Syst Video Technol 32(10):6700\u20136713. https:\/\/doi.org\/10.1109\/TCSVT.2022.3168279","journal-title":"IEEE Trans Circ Syst Video Technol"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01986-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-025-01986-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-01986-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,7]],"date-time":"2025-09-07T00:06:59Z","timestamp":1757203619000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-025-01986-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,1]]},"references-count":62,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["1986"],"URL":"https:\/\/doi.org\/10.1007\/s40747-025-01986-7","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,7,1]]},"assertion":[{"value":"2 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 May 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 July 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"367"}}