{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,20]],"date-time":"2025-09-20T04:30:48Z","timestamp":1758342648978,"version":"3.44.0"},"reference-count":55,"publisher":"Springer Science and Business Media LLC","issue":"7","license":[{"start":{"date-parts":[[2025,4,14]],"date-time":"2025-04-14T00:00:00Z","timestamp":1744588800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,4,14]],"date-time":"2025-04-14T00:00:00Z","timestamp":1744588800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62373102"],"award-info":[{"award-number":["62373102"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2025,5]]},"DOI":"10.1007\/s10489-025-06470-w","type":"journal-article","created":{"date-parts":[[2025,4,14]],"date-time":"2025-04-14T00:30:21Z","timestamp":1744590621000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["MKDFusion: modality knowledge decoupled for infrared and visible image fusion"],"prefix":"10.1007","volume":"55","author":[{"given":"Yucheng","family":"Zhang","sequence":"first","affiliation":[]},{"given":"You","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Lin","family":"Chai","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,14]]},"reference":[{"key":"6470_CR1","doi-asserted-by":"publisher","first-page":"108635","DOI":"10.1016\/j.compbiomed.2024.108635","volume":"177","author":"Y Li","year":"2024","unstructured":"Li Y, El Habib Daho M, Conze P-H, Zeghlache R, Le Boit\u00e9 H, Tadayoni R, Cochener B, Lamard M, Quellec G (2024) A review of deep learning-based information fusion techniques for multimodal medical image classification. Comput Biol Med 177:108635. https:\/\/doi.org\/10.1016\/j.compbiomed.2024.108635","journal-title":"Comput Biol Med"},{"key":"6470_CR2","doi-asserted-by":"publisher","first-page":"100327","DOI":"10.1016\/j.dajour.2023.100327","volume":"9","author":"S Kalamkar","year":"2023","unstructured":"Kalamkar S, Geetha MA (2023) Multimodal image fusion: A systematic review. Decision Anal J 9:100327. https:\/\/doi.org\/10.1016\/j.dajour.2023.100327","journal-title":"Decision Anal J"},{"key":"6470_CR3","doi-asserted-by":"publisher","first-page":"104179","DOI":"10.1016\/j.jvcir.2024.104179","volume":"101","author":"K Yang","year":"2024","unstructured":"Yang K, Xiang W, Chen Z, Zhang J, Liu Y (2024) A review on infrared and visible image fusion algorithms based on neural networks. J Vis Commun Image Represent 101:104179. https:\/\/doi.org\/10.1016\/j.jvcir.2024.104179","journal-title":"J Vis Commun Image Represent"},{"key":"6470_CR4","doi-asserted-by":"publisher","first-page":"108463","DOI":"10.1016\/j.compbiomed.2024.108463","volume":"174","author":"Y Song","year":"2024","unstructured":"Song Y, Dai Y, Liu W, Liu Y, Liu X, Yu Q, Liu X, Que N, Li M (2024) DesTrans: A medical image fusion method based on transformer and improved DenseNet. Comput Biol Med 174:108463. https:\/\/doi.org\/10.1016\/j.compbiomed.2024.108463","journal-title":"Comput Biol Med"},{"key":"6470_CR5","doi-asserted-by":"publisher","first-page":"112159","DOI":"10.1016\/j.knosys.2024.112159","volume":"300","author":"W Cheng","year":"2024","unstructured":"Cheng W, Feng Y, Song L, Wang X (2024) DMF2Net: Dynamic multi-level feature fusion network for heterogeneous remote sensing image change detection. Knowl-Based Syst 300:112159. https:\/\/doi.org\/10.1016\/j.knosys.2024.112159","journal-title":"Knowl-Based Syst"},{"key":"6470_CR6","doi-asserted-by":"publisher","unstructured":"Wang D, Liu J, Liu R, Fan X (2023) An interactively reinforced paradigm for joint infrared-visible image fusion and saliency object detection. Inf Fusion. 98:101828. https:\/\/doi.org\/10.1016\/j.inffus.2023.101828","DOI":"10.1016\/j.inffus.2023.101828"},{"key":"6470_CR7","doi-asserted-by":"publisher","first-page":"105343","DOI":"10.1016\/j.infrared.2024.105343","volume":"139","author":"R An","year":"2024","unstructured":"An R, Liu G, Qian Y, Xing M, Tang H (2024) MRASFusion: A multi-scale residual attention infrared and visible image fusion network based on semantic segmentation guidance. Infrared Phys Technol 139:105343. https:\/\/doi.org\/10.1016\/j.infrared.2024.105343","journal-title":"Infrared Phys Technol"},{"key":"6470_CR8","doi-asserted-by":"publisher","unstructured":"Tlig M, Bouchouicha M, Sayadi M, Moreau E (2022) Visible and Infrared Image Fusion Framework for Fire Semantic Segmentation Using U-Net-ResNet50. In: 2022 IEEE Information Technologies & Smart Industrial Systems (ITSIS). IEEE. https:\/\/doi.org\/10.1109\/itsis56166.2022.10118361. Accessed August 11, 2024","DOI":"10.1109\/itsis56166.2022.10118361"},{"key":"6470_CR9","doi-asserted-by":"publisher","first-page":"6480","DOI":"10.1364\/ao.55.006480","volume":"55","author":"Z Zhou","year":"2016","unstructured":"Zhou Z, Dong M, Xie X, Gao Z (2016) Fusion of infrared and visible images for night-vision context enhancement. Appl Opt 55:6480. https:\/\/doi.org\/10.1364\/ao.55.006480","journal-title":"Appl Opt"},{"key":"6470_CR10","doi-asserted-by":"publisher","unstructured":"Sun W, Hu S, Liu S, Sun Y (2014) Infrared and visible image fusion based on object extraction and adaptive pulse coupled neural network via non-subsampled Shearlet transform. In: 2014 12th International Conference on Signal Processing (ICSP), IEEE pp 946\u2013951. https:\/\/doi.org\/10.1109\/icosp.2014.7015144. Accessed December 20, 2024","DOI":"10.1109\/icosp.2014.7015144."},{"key":"6470_CR11","doi-asserted-by":"publisher","unstructured":"Zhang C (2020) Multi-focus Image Fusion Based on Convolutional Sparse Representation with Mask Simulation. In: Smart Innovation, Systems and Technologies, Springer Singapore, Singapore, pp 159\u2013168. https:\/\/doi.org\/10.1007\/978-981-15-3867-4_19. Accessed August 11, 2024","DOI":"10.1007\/978-981-15-3867-4_19"},{"key":"6470_CR12","doi-asserted-by":"publisher","unstructured":"J. Mou W, Gao Z (2013) Song, Image fusion based on non-negative matrix factorization and infrared feature extraction. In: 2013 6th International Congress on Image and Signal Processing (CISP). IEEE. https:\/\/doi.org\/10.1109\/cisp.2013.6745210. Accessed August 11, 2024","DOI":"10.1109\/cisp.2013.6745210."},{"key":"6470_CR13","doi-asserted-by":"publisher","unstructured":"Tan X, Guo L (2020) Visible and infrared image fusion based on visual saliency detection. In: 2020 19th International Symposium on Distributed Computing and Applications for Business Engineering and Science (DCABES). IEEE. https:\/\/doi.org\/10.1109\/dcabes50732.2020.00043. Accessed August 11, 2024","DOI":"10.1109\/dcabes50732.2020.00043"},{"key":"6470_CR14","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li H, Wu X-J, Kittler J (2021) RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf Fusion 73:72\u201386. https:\/\/doi.org\/10.1016\/j.inffus.2021.02.023","journal-title":"Inf Fusion"},{"key":"6470_CR15","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/tip.2018.2887342","volume":"28","author":"H Li","year":"2019","unstructured":"Li H, Wu X-J (2019) DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans Image Process 28:2614\u20132623. https:\/\/doi.org\/10.1109\/tip.2018.2887342","journal-title":"IEEE Trans Image Process"},{"key":"6470_CR16","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang Y, Liu Y, Sun P, Yan H, Zhao X, Zhang L (2020) IFCNN: A general image fusion framework based on convolutional neural network. Inf Fusion 54:99\u2013118. https:\/\/doi.org\/10.1016\/j.inffus.2019.07.011","journal-title":"Inf Fusion"},{"key":"6470_CR17","doi-asserted-by":"publisher","first-page":"477","DOI":"10.1016\/j.inffus.2022.10.034","volume":"91","author":"L Tang","year":"2023","unstructured":"Tang L, Xiang X, Zhang H, Gong M, Ma J (2023) DIVFusion: Darkness-free infrared and visible image fusion. Inf Fusion 91:477\u2013493. https:\/\/doi.org\/10.1016\/j.inffus.2022.10.034","journal-title":"Inf Fusion"},{"key":"6470_CR18","doi-asserted-by":"publisher","first-page":"11373","DOI":"10.1007\/s10489-024-05722-5","volume":"54","author":"S Liu","year":"2024","unstructured":"Liu S, Liu Y, Su Y, Zhang Y (2024) EDOM-MFIF: an end-to-end decision optimization model for multi-focus image fusion. Appl Intell 54:11373\u201311399. https:\/\/doi.org\/10.1007\/s10489-024-05722-5","journal-title":"Appl Intell"},{"key":"6470_CR19","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma J, Yu W, Liang P, Li C, Jiang J (2019) FusionGAN: A generative adversarial network for infrared and visible image fusion. Information Fusion 48:11\u201326. https:\/\/doi.org\/10.1016\/j.inffus.2018.09.004","journal-title":"Information Fusion"},{"key":"6470_CR20","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/tim.2020.3038013","volume":"70","author":"J Ma","year":"2021","unstructured":"Ma J, Zhang H, Shao Z, Liang P, Xu H (2021) GANMcC: A generative adversarial network with multiclassification constraints for infrared and visible image fusion. IEEE Trans Instrum Meas 70:1\u201314. https:\/\/doi.org\/10.1109\/tim.2020.3038013","journal-title":"IEEE Trans Instrum Meas"},{"key":"6470_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2022.104522","volume":"128","author":"X Yang","year":"2023","unstructured":"Yang X, Huo H, Wang R, Li C, Liu X, Li J (2023) DGLT-Fusion: A decoupled global\u2013local infrared and visible image fusion transformer. Infrared Phys Technol 128:104522. https:\/\/doi.org\/10.1016\/j.infrared.2022.104522","journal-title":"Infrared Phys Technol"},{"key":"6470_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2023.117058","volume":"119","author":"L Karacan","year":"2023","unstructured":"Karacan L (2023) Multi-image transformer for multi-focus image fusion. Signal Process Image Commun 119:117058. https:\/\/doi.org\/10.1016\/j.image.2023.117058","journal-title":"Signal Process Image Commun"},{"key":"6470_CR23","doi-asserted-by":"publisher","first-page":"4980","DOI":"10.1109\/tip.2020.2977573","volume":"29","author":"J Ma","year":"2020","unstructured":"Ma J, Xu H, Jiang J, Mei X, Zhang X-P (2020) DDcGAN: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans Image Process 29:4980\u20134995. https:\/\/doi.org\/10.1109\/tip.2020.2977573","journal-title":"IEEE Trans Image Process"},{"key":"6470_CR24","doi-asserted-by":"publisher","unstructured":"Wu A, Liu R, Han Y, Zhu L, YangY (2021) Vector-Decomposed Disentanglement for Domain-Invariant Object Detection. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV). IEEE. https:\/\/doi.org\/10.1109\/iccv48922.2021.00921. Accessed August 11, 2024","DOI":"10.1109\/iccv48922.2021.00921."},{"key":"6470_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/tim.2023.3237814","volume":"72","author":"J Wang","year":"2023","unstructured":"Wang J, Xi X, Li D, Li F (2023) FusionGRAM: An infrared and visible image fusion framework based on gradient residual and attention mechanism. IEEE Trans Instrum Meas 72:1\u201312. https:\/\/doi.org\/10.1109\/tim.2023.3237814","journal-title":"IEEE Trans Instrum Meas"},{"key":"6470_CR26","doi-asserted-by":"publisher","first-page":"103407","DOI":"10.1016\/j.cviu.2022.103407","volume":"218","author":"H Xu","year":"2022","unstructured":"Xu H, Gong M, Tian X, Huang J, Ma J (2022) CUFD: An encoder\u2013decoder network for visible and infrared image fusion based on common and unique feature decomposition. Comput Vis Image Underst 218:103407. https:\/\/doi.org\/10.1016\/j.cviu.2022.103407","journal-title":"Comput Vis Image Underst"},{"key":"6470_CR27","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","volume":"83\u201384","author":"L Tang","year":"2022","unstructured":"Tang L, Yuan J, Zhang H, Jiang X, Ma J (2022) PIAFusion: A progressive infrared and visible image fusion network based on illumination aware. Inf Fusion 83\u201384:79\u201392. https:\/\/doi.org\/10.1016\/j.inffus.2022.03.007","journal-title":"Inf Fusion"},{"key":"6470_CR28","doi-asserted-by":"publisher","first-page":"640","DOI":"10.1109\/tci.2020.2965304","volume":"6","author":"R Hou","year":"2020","unstructured":"Hou R, Zhou D, Nie R, Liu D, Xiong L, Guo Y, Yu C (2020) VIF-Net: an unsupervised framework for infrared and visible image fusion. IEEE Trans Comput Imaging 6:640\u2013651. https:\/\/doi.org\/10.1109\/tci.2020.2965304","journal-title":"IEEE Trans Comput Imaging"},{"key":"6470_CR29","doi-asserted-by":"publisher","first-page":"104957","DOI":"10.1016\/j.imavis.2024.104957","volume":"144","author":"J Huang","year":"2024","unstructured":"Huang J, Chen Z, Ma Y, Fan F, Tang L, Xiang X (2024) PTET: A progressive token exchanging transformer for infrared and visible image fusion. Image Vis Comput 144:104957. https:\/\/doi.org\/10.1016\/j.imavis.2024.104957","journal-title":"Image Vis Comput"},{"key":"6470_CR30","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2024.111658","volume":"293","author":"L Mei","year":"2024","unstructured":"Mei L, Hu X, Ye Z, Tang L, Wang Y, Li D, Liu Y, Hao X, Lei C, Xu C, Yang W (2024) GTMFuse: Group-attention transformer-driven multiscale dense feature-enhanced network for infrared and visible image fusion. Knowl-Based Syst 293:111658. https:\/\/doi.org\/10.1016\/j.knosys.2024.111658","journal-title":"Knowl-Based Syst"},{"key":"6470_CR31","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1016\/j.neucom.2023.01.033","volume":"527","author":"J Chen","year":"2023","unstructured":"Chen J, Ding J, Yu Y, Gong W (2023) THFuse: An infrared and visible image fusion network using transformer and hybrid feature extractor. Neurocomputing 527:71\u201382. https:\/\/doi.org\/10.1016\/j.neucom.2023.01.033","journal-title":"Neurocomputing"},{"key":"6470_CR32","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.109295","volume":"137","author":"W Tang","year":"2023","unstructured":"Tang W, He F, Liu Y (2023) TCCFusion: An infrared and visible image fusion method based on transformer and cross correlation. Pattern Recogn 137:109295. https:\/\/doi.org\/10.1016\/j.patcog.2022.109295","journal-title":"Pattern Recogn"},{"key":"6470_CR33","doi-asserted-by":"publisher","unstructured":"Zhao Z, Bai H, Zhang J, Zhang Y, Xu S, Lin Z, Timofte R, Van Gool L (2023) CDDFuse: Correlation-Driven Dual-Branch Feature Decomposition for Multi-Modality Image Fusion. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE. https:\/\/doi.org\/10.1109\/cvpr52729.2023.00572. Accessed August 11, 2024","DOI":"10.1109\/cvpr52729.2023.00572"},{"key":"6470_CR34","doi-asserted-by":"publisher","first-page":"1383","DOI":"10.1109\/tmm.2020.2997127","volume":"23","author":"J Li","year":"2021","unstructured":"Li J, Huo H, Li C, Wang R, Feng Q (2021) AttentionFGAN: Infrared and visible image fusion using attention-based generative adversarial networks. IEEE Trans Multimedia 23:1383\u20131396. https:\/\/doi.org\/10.1109\/tmm.2020.2997127","journal-title":"IEEE Trans Multimedia"},{"key":"6470_CR35","unstructured":"Gu A, Dao T (2023) Mamba: Linear-Time Sequence Modeling with Selective State Spaces. arXiv.Org. https:\/\/arxiv.org\/abs\/2312.00752"},{"key":"6470_CR36","doi-asserted-by":"publisher","first-page":"128104","DOI":"10.1016\/j.neucom.2024.128104","volume":"599","author":"Y Ge","year":"2024","unstructured":"Ge Y, Chen Z, Yu M, Yue Q, You R, Zhu L (2024) MambaTSR: You only need 90k parameters for traffic sign recognition. Neurocomputing 599:128104. https:\/\/doi.org\/10.1016\/j.neucom.2024.128104","journal-title":"Neurocomputing"},{"key":"6470_CR37","doi-asserted-by":"publisher","first-page":"112203","DOI":"10.1016\/j.knosys.2024.112203","volume":"300","author":"C Ma","year":"2024","unstructured":"Ma C, Wang Z (2024) Semi-Mamba-UNet: Pixel-level contrastive and cross-supervised visual Mamba-based UNet for semi-supervised medical image segmentation. Knowl-Based Syst 300:112203. https:\/\/doi.org\/10.1016\/j.knosys.2024.112203","journal-title":"Knowl-Based Syst"},{"key":"6470_CR38","doi-asserted-by":"crossref","unstructured":"Xie X, Cui Y, Leong C-I, Tan T, Zhang X, Zheng X (2024) FusionMamba: Dynamic Feature Enhancement for Multimodal Image Fusion with Mamba. arXiv.Org. https:\/\/arxiv.org\/html\/2404.09498v2","DOI":"10.1007\/s44267-024-00072-9"},{"key":"6470_CR39","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/tim.2021.3056645","volume":"70","author":"H Xu","year":"2021","unstructured":"Xu H, Wang X, Ma J (2021) DRF: Disentangled representation for visible and infrared image fusion. IEEE Trans Instrum Meas 70:1\u201313. https:\/\/doi.org\/10.1109\/tim.2021.3056645","journal-title":"IEEE Trans Instrum Meas"},{"key":"6470_CR40","unstructured":"Liu Y, Tian Y, Yu H, Xie L, Wang Y, Ye Q, Liu Y (2024) VMamba: Visual State Space Model, arXiv.Org. https:\/\/arxiv.org\/html\/2401.10166v1"},{"key":"6470_CR41","unstructured":"Ganin Y, Lempitsky V (2014) Unsupervised Domain Adaptation by Backpropagation, arXiv.Org. https:\/\/arxiv.org\/abs\/1409.7495"},{"key":"6470_CR42","doi-asserted-by":"publisher","first-page":"10118","DOI":"10.1109\/tits.2023.3268063","volume":"24","author":"G Li","year":"2023","unstructured":"Li G, Qian X, Qu X (2023) SOSMaskFuse: An infrared and visible image fusion architecture based on salient object segmentation mask. IEEE Trans Intell Transp Syst 24:10118\u201310137. https:\/\/doi.org\/10.1109\/tits.2023.3268063","journal-title":"IEEE Trans Intell Transp Syst"},{"key":"6470_CR43","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/tim.2022.3149101","volume":"71","author":"W Xiao","year":"2022","unstructured":"Xiao W, Zhang Y, Wang H, Li F, Jin H (2022) Heterogeneous knowledge distillation for simultaneous infrared-visible image fusion and super-resolution. IEEE Trans Instrum Meas 71:1\u201315. https:\/\/doi.org\/10.1109\/tim.2022.3149101","journal-title":"IEEE Trans Instrum Meas"},{"issue":"5","key":"6470_CR44","first-page":"484","volume":"4","author":"M Deshmukh","year":"2010","unstructured":"Deshmukh M, Bhosale U (2010) Image fusion and image quality assessment of fused images. Int J Image Process 4(5):484","journal-title":"Int J Image Process"},{"key":"6470_CR45","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1016\/j.dib.2017.09.038","volume":"15","author":"A Toet","year":"2017","unstructured":"Toet A (2017) The TNO multiband image data collection. Data Brief 15:249\u2013251. https:\/\/doi.org\/10.1016\/j.dib.2017.09.038","journal-title":"Data Brief"},{"key":"6470_CR46","doi-asserted-by":"publisher","first-page":"12484","DOI":"10.1609\/aaai.v34i07.6936","volume":"34","author":"H Xu","year":"2020","unstructured":"Xu H, Ma J, Le Z, Jiang J, Guo X (2020) FusionDN: A unified densely connected network for image fusion. Proc AAAI Conf Art Intell 34:12484\u201312491. https:\/\/doi.org\/10.1609\/aaai.v34i07.6936","journal-title":"Proc AAAI Conf Art Intell"},{"key":"6470_CR47","doi-asserted-by":"publisher","unstructured":"Ha Q, Watanabe K, Karasawa T, Ushiku Y, Harada T (2017) MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE. https:\/\/doi.org\/10.1109\/iros.2017.8206396. Accessed August 11, 2024","DOI":"10.1109\/iros.2017.8206396"},{"key":"6470_CR48","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/tpami.2020.3012548","volume":"44","author":"H Xu","year":"2022","unstructured":"Xu H, Ma J, Jiang J, Guo X, Ling H (2022) U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans Pattern Anal Mach Intell 44:502\u2013518. https:\/\/doi.org\/10.1109\/tpami.2020.3012548","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"6470_CR49","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang L, Yuan J, Ma J (2022) Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Information Fusion 82:28\u201342. https:\/\/doi.org\/10.1016\/j.inffus.2021.12.004","journal-title":"Information Fusion"},{"key":"6470_CR50","doi-asserted-by":"publisher","first-page":"102147","DOI":"10.1016\/j.inffus.2023.102147","volume":"103","author":"H Li","year":"2024","unstructured":"Li H, Wu X-J (2024) CrossFuse: A novel cross attention mechanism based infrared and visible image fusion approach. Information Fusion 103:102147. https:\/\/doi.org\/10.1016\/j.inffus.2023.102147","journal-title":"Information Fusion"},{"key":"6470_CR51","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/tim.2022.3191664","volume":"71","author":"Z Wang","year":"2022","unstructured":"Wang Z, Chen Y, Shao W, Li H, Zhang L (2022) SwinFuse: A residual swin transformer fusion network for infrared and visible images. IEEE Trans Instrum Meas 71:1\u201312. https:\/\/doi.org\/10.1109\/tim.2022.3191664","journal-title":"IEEE Trans Instrum Meas"},{"key":"6470_CR52","doi-asserted-by":"publisher","unstructured":"Qi W, Zhang Z, Wang Z (2024) DMFuse: Diffusion model guided cross-attention learning for infrared and visible image fusion. Chin J Inf Fusion 1, 226\u2013241. https:\/\/doi.org\/10.62762\/cjif.2024.655617","DOI":"10.62762\/cjif.2024.655617"},{"key":"6470_CR53","unstructured":"Harvard medical website. http:\/\/www.med.harvard.edu\/AANLIB"},{"key":"6470_CR54","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1016\/j.inffus.2021.06.001","volume":"76","author":"H Xu","year":"2021","unstructured":"Xu H, Ma J (2021) EMFusion: An unsupervised enhanced medical image fusion network. Information Fusion 76:177\u2013186. https:\/\/doi.org\/10.1016\/j.inffus.2021.06.001","journal-title":"Information Fusion"},{"key":"6470_CR55","doi-asserted-by":"publisher","first-page":"5134","DOI":"10.1109\/tip.2022.3193288","volume":"31","author":"W Tang","year":"2022","unstructured":"Tang W, He F, Liu Y, Duan Y (2022) MATR: Multimodal medical image fusion via multiscale adaptive transformer. IEEE Trans Image Process 31:5134\u20135149. https:\/\/doi.org\/10.1109\/tip.2022.3193288","journal-title":"IEEE Trans Image Process"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-025-06470-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-025-06470-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-025-06470-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,19]],"date-time":"2025-09-19T19:37:56Z","timestamp":1758310676000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-025-06470-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,14]]},"references-count":55,"journal-issue":{"issue":"7","published-print":{"date-parts":[[2025,5]]}},"alternative-id":["6470"],"URL":"https:\/\/doi.org\/10.1007\/s10489-025-06470-w","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"type":"print","value":"0924-669X"},{"type":"electronic","value":"1573-7497"}],"subject":[],"published":{"date-parts":[[2025,4,14]]},"assertion":[{"value":"14 March 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 April 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interest"}}],"article-number":"653"}}