{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T08:56:00Z","timestamp":1770800160228,"version":"3.50.0"},"reference-count":60,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T00:00:00Z","timestamp":1764806400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T00:00:00Z","timestamp":1764806400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"name":"Independent Research Project , Naval University of Engineering, PLA China","award":["No.2025500130"],"award-info":[{"award-number":["No.2025500130"]}]},{"name":"Jiangxi Provincial Natural Science Foundation of China","award":["No.20242BAB25064"],"award-info":[{"award-number":["No.20242BAB25064"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s00530-025-02080-1","type":"journal-article","created":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T11:33:59Z","timestamp":1764848039000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["MCDFus: visible and infrared image fusion via Mamba-CNN for enhanced UAV perception"],"prefix":"10.1007","volume":"32","author":[{"given":"Yu","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Hao","family":"Cha","sequence":"additional","affiliation":[]},{"given":"HanRui","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"XiongHui","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zhuo","family":"Chen","sequence":"additional","affiliation":[]},{"given":"JiaMing","family":"Bai","sequence":"additional","affiliation":[]},{"given":"TingTing","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Yong Cheng","family":"Du","sequence":"additional","affiliation":[]},{"given":"Xian","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,4]]},"reference":[{"issue":"1","key":"2080_CR1","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1007\/s11370-022-00452-4","volume":"16","author":"SAH Mohsan","year":"2023","unstructured":"Mohsan, S.A.H., Othman, N.Q.H., Li, Y., Alsharif, M.H., Khan, M.A.: Unmanned aerial vehicles (UAVS): practical aspects, applications, open challenges, security issues, and future trends. Intel. Serv. Robot. 16(1), 109\u2013137 (2023)","journal-title":"Intel. Serv. Robot."},{"issue":"8","key":"2080_CR2","doi-asserted-by":"publisher","first-page":"10535","DOI":"10.1109\/TPAMI.2023.3261282","volume":"45","author":"X Zhang","year":"2023","unstructured":"Zhang, X., Demiris, Y.: Visible and infrared image fusion using deep learning. IEEE Trans. Pattern Anal. Mach. Intell. 45(8), 10535\u201310554 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2080_CR3","doi-asserted-by":"crossref","unstructured":"Gao, Z., Li, D., Kuai, Y., Chen, R., Wen, G.: Visible-infrared image alignment for unmanned aerial vehicles: benchmark and new baseline. IEEE Trans. Geosci. Remote Sens. (2025)","DOI":"10.1109\/TGRS.2025.3528634"},{"key":"2080_CR4","volume":"115","author":"Z Ma","year":"2022","unstructured":"Ma, Z., Chen, J.: Adaptive path planning method for UAVS in complex environments. Int. J. Appl. Earth Obs. Geoinf. 115, 103133 (2022)","journal-title":"Int. J. Appl. Earth Obs. Geoinf."},{"key":"2080_CR5","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mao, H., Wu, C.-Y., Feichtenhofer, C., Darrell, T., Xie, S.: A convnet for the 2020s. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11976\u201311986 (2022)","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"2080_CR6","doi-asserted-by":"publisher","first-page":"128","DOI":"10.1016\/j.inffus.2020.11.009","volume":"69","author":"Y Long","year":"2021","unstructured":"Long, Y., Jia, H., Zhong, Y., Jiang, Y., Jia, Y.: Rxdnfuse: a aggregated residual dense network for infrared and visible image fusion. Inf. Fusion. 69, 128\u2013141 (2021)","journal-title":"Inf. Fusion."},{"key":"2080_CR7","first-page":"1","volume":"70","author":"J Ma","year":"2021","unstructured":"Ma, J., Tang, L., Xu, M., Zhang, H., Xiao, G.: Stdfusionnet: an infrared and visible image fusion network based on salient target detection. IEEE Trans. Instrum. Meas. 70, 1\u201313 (2021)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"2080_CR8","doi-asserted-by":"crossref","unstructured":"Zhao, W., Xie, S., Zhao, F., He, Y., Lu, H.: Metafusion: infrared and visible image fusion via meta-feature embedding from object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13955\u201313965 (2023)","DOI":"10.1109\/CVPR52729.2023.01341"},{"key":"2080_CR9","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2020.3022438","volume":"70","author":"L Jian","year":"2020","unstructured":"Jian, L., Yang, X., Liu, Z., Jeon, G., Gao, M., Chisholm, D.: Sedrfuse: a symmetric encoder-decoder with residual block network for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 70, 1\u201315 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"1","key":"2080_CR10","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1109\/TCSVT.2021.3056725","volume":"32","author":"J Liu","year":"2021","unstructured":"Liu, J., Fan, X., Jiang, J., Liu, R., Luo, Z.: Learning a deep multi-scale feature ensemble and an edge-attention guidance for image fusion. IEEE Trans. Circ. Syst. Video Technol. 32(1), 105\u2013119 (2021)","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"key":"2080_CR11","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Wu, X.-J., Kittler, J.: Rfn-nest: an end-to-end residual fusion network for infrared and visible images. Inf. Fusion 73, 72\u201386 (2021)","journal-title":"Inf. Fusion"},{"key":"2080_CR12","doi-asserted-by":"publisher","first-page":"8988","DOI":"10.1109\/TMM.2023.3243659","volume":"25","author":"J Zhang","year":"2023","unstructured":"Zhang, J., Jiao, L., Ma, W., Liu, F., Liu, X., Li, L., Chen, P., Yang, S.: Transformer based conditional gan for multimodal image fusion. IEEE Trans. Multimedia 25, 8988\u20139001 (2023)","journal-title":"IEEE Trans. Multimedia"},{"issue":"5","key":"2080_CR13","doi-asserted-by":"publisher","first-page":"288","DOI":"10.1007\/s00530-024-01500-y","volume":"30","author":"B Liu","year":"2024","unstructured":"Liu, B., Wu, Z., Li, X.: Msadrcn: meta-learning based joint super-resolution fusion of infrared and visible images. Multimed. Syst. 30(5), 288 (2024)","journal-title":"Multimed. Syst."},{"key":"2080_CR14","doi-asserted-by":"crossref","unstructured":"Zou, X., Tang, J.: Guided fusion of infrared and visible images using gradient-based attentive generative adversarial networks. Visual Comput. 1\u201318 (2025)","DOI":"10.1007\/s00371-025-03801-3"},{"key":"2080_CR15","doi-asserted-by":"publisher","first-page":"5413","DOI":"10.1109\/TMM.2022.3192661","volume":"25","author":"W Tang","year":"2022","unstructured":"Tang, W., He, F., Liu, Y.: Ydtr: infrared and visible image fusion via y-shape dynamic transformer. IEEE Trans. Multimed. 25, 5413\u20135428 (2022)","journal-title":"IEEE Trans. Multimed."},{"key":"2080_CR16","doi-asserted-by":"crossref","unstructured":"Rao, D., Xu, T., Wu, X.-J.: Tgfuse: an infrared and visible image fusion approach based on transformer and generative adversarial network. IEEE Trans. Image Process. (2023)","DOI":"10.1109\/TIP.2023.3273451"},{"issue":"2","key":"2080_CR17","doi-asserted-by":"publisher","first-page":"174","DOI":"10.1007\/s00530-025-01763-z","volume":"31","author":"L Huang","year":"2025","unstructured":"Huang, L., Xue, Z., Ning, Q., Guo, Y., Li, Y.: A guidance and alignment transformer model for visible-infrared person re-identification. Multimed. Syst. 31(2), 174 (2025)","journal-title":"Multimed. Syst."},{"issue":"03","key":"2080_CR18","doi-asserted-by":"publisher","first-page":"1850018","DOI":"10.1142\/S0219691318500182","volume":"16","author":"Y Liu","year":"2018","unstructured":"Liu, Y., Chen, X., Cheng, J., Peng, H., Wang, Z.: Infrared and visible image fusion with convolutional neural networks. Int. J. Wavelets Multiresolut. Inf. Process. 16(03), 1850018 (2018)","journal-title":"Int. J. Wavelets Multiresolut. Inf. Process."},{"key":"2080_CR19","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: Fusiongan: a generative adversarial network for infrared and visible image fusion. Inf. Fusion 48, 11\u201326 (2019)","journal-title":"Inf. Fusion"},{"key":"2080_CR20","doi-asserted-by":"publisher","first-page":"5705","DOI":"10.1109\/TIP.2023.3322046","volume":"32","author":"J Yue","year":"2023","unstructured":"Yue, J., Fang, L., Xia, S., Deng, Y., Ma, J.: Diffusion: toward high color fidelity in infrared and visible image fusion with diffusion models. IEEE Trans. Image Process. 32, 5705\u20135720 (2023)","journal-title":"IEEE Trans. Image Process."},{"key":"2080_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2025.105834","volume":"148","author":"Y Chen","year":"2025","unstructured":"Chen, Y., Liu, G., Xing, M., Li, K., Xiao, G.: Denoising diffusion based infrared and visible image fusion with transformer. Infrared Phys. Technol. 148, 105834 (2025)","journal-title":"Infrared Phys. Technol."},{"key":"2080_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102450","volume":"110","author":"X Yi","year":"2024","unstructured":"Yi, X., Tang, L., Zhang, H., Xu, H., Ma, J.: Diff-if: multi-modality image fusion via diffusion model with fusion knowledge prior. Inf. Fusion 110, 102450 (2024)","journal-title":"Inf. Fusion"},{"key":"2080_CR23","unstructured":"Wang, X., Guan, Z., Qian, W., Cao, J., Wang, C., Ma, R.: Stfuse: infrared and visible image fusion via semisupervised transfer learning. IEEE Trans. Neural Netw. Learn. Syst. (2023)"},{"issue":"4","key":"2080_CR24","doi-asserted-by":"publisher","first-page":"939","DOI":"10.3390\/rs14040939","volume":"14","author":"JA Bhutto","year":"2022","unstructured":"Bhutto, J.A., Tian, L., Du, Q., Sun, Z., Yu, L., Soomro, T.A.: An improved infrared and visible image fusion using an adaptive contrast enhancement method and deep learning network with transfer learning. Remote Sens. 14(4), 939 (2022)","journal-title":"Remote Sens."},{"key":"2080_CR25","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2020.103237","volume":"105","author":"Z Gao","year":"2020","unstructured":"Gao, Z., Zhang, Y., Li, Y.: Extracting features from infrared images using convolutional neural networks and transfer learning. Infrared Phys. Technol. 105, 103237 (2020)","journal-title":"Infrared Phys. Technol."},{"key":"2080_CR26","first-page":"1","volume":"71","author":"Z Zhu","year":"2022","unstructured":"Zhu, Z., Yang, X., Lu, R., Shen, T., Xie, X., Zhang, T.: Clf-net: contrastive learning for infrared and visible image fusion network. IEEE Trans. Instrum. Meas. 71, 1\u201315 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"5","key":"2080_CR27","doi-asserted-by":"publisher","first-page":"1748","DOI":"10.1007\/s11263-023-01952-1","volume":"132","author":"J Liu","year":"2024","unstructured":"Liu, J., Lin, R., Wu, G., Liu, R., Luo, Z., Fan, X.: Coconet: coupled contrastive learning network with multi-level feature ensemble for multi-modality image fusion. Int. J. Comput. Vision 132(5), 1748\u20131775 (2024)","journal-title":"Int. J. Comput. Vision"},{"key":"2080_CR28","doi-asserted-by":"crossref","unstructured":"Gupta, A.K., Barnwal, M., Mishra, D.: A contrastive learning approach for infrared-visible image fusion. In: International Conference on Pattern Recognition and Machine Intelligence, pp. 199\u2013208 (2023). Springer","DOI":"10.1007\/978-3-031-45170-6_21"},{"key":"2080_CR29","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Bai, H., Zhu, Y., Zhang, J., Xu, S., Zhang, Y., Zhang, K., Meng, D., Timofte, R., Van\u00a0Gool, L.: Ddfm: denoising diffusion model for multi-modality image fusion. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 8082\u20138093 (2023)","DOI":"10.1109\/ICCV51070.2023.00742"},{"key":"2080_CR30","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.102039","volume":"102","author":"X Wang","year":"2024","unstructured":"Wang, X., Guan, Z., Qian, W., Cao, J., Liang, S., Yan, J.: Cs2fusion: contrastive learning for self-supervised infrared and visible image fusion by estimating feature compensation map. Inf. Fusion 102, 102039 (2024)","journal-title":"Inf. Fusion"},{"key":"2080_CR31","unstructured":"Gu, A., Dao, T.: Mamba: linear-time sequence modeling with selective state spaces. arXiv preprint arXiv:2312.00752 (2023)"},{"key":"2080_CR32","unstructured":"Zhu, L., Liao, B., Zhang, Q., Wang, X., Liu, W., Wang, X.: Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model (2024). https:\/\/arxiv.org\/abs\/2401.09417"},{"key":"2080_CR33","doi-asserted-by":"crossref","unstructured":"Xie, X., Cui, Y., Tan, T., Zheng, X., Yu, Z.: FusionMamba: Dynamic Feature Enhancement for Multimodal Image Fusion with Mamba (2025). https:\/\/arxiv.org\/abs\/2404.09498","DOI":"10.1007\/s44267-024-00072-9"},{"key":"2080_CR34","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/tgrs.2024.3496073","volume":"62","author":"S Peng","year":"2024","unstructured":"Peng, S., Zhu, X., Deng, H., Deng, L.-J., Lei, Z.: Fusionmamba: efficient remote sensing image fusion with state space model. IEEE Trans. Geosci. Remote Sens. 62, 1\u201316 (2024). https:\/\/doi.org\/10.1109\/tgrs.2024.3496073","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"2080_CR35","unstructured":"Li, Z., Pan, H., Zhang, K., Wang, Y., Yu, F.: MambaDFuse: A Mamba-Based Dual-phase Model for Multi-Modality Image Fusion (2024). https:\/\/arxiv.org\/abs\/2404.08406"},{"issue":"1","key":"2080_CR36","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1007\/s44267-023-00030-x","volume":"1","author":"B Du","year":"2023","unstructured":"Du, B., Du, C., Yu, L.: Megf-net: multi-exposure generation and fusion network for vehicle detection under dim light conditions. Visual Intell. 1(1), 28 (2023)","journal-title":"Visual Intell."},{"issue":"1","key":"2080_CR37","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1007\/s44267-024-00040-3","volume":"2","author":"Z Jia","year":"2024","unstructured":"Jia, Z., Sun, S., Liu, G., Liu, B.: Mssd: multi-scale self-distillation for object detection. Visual Intell. 2(1), 8 (2024)","journal-title":"Visual Intell."},{"issue":"1","key":"2080_CR38","doi-asserted-by":"publisher","first-page":"6","DOI":"10.1007\/s44267-024-00039-w","volume":"2","author":"L Fang","year":"2024","unstructured":"Fang, L., Wang, Q., Ye, L.: Glgnet: light field angular super resolution with arbitrary interpolation rates. Visual Intell. 2(1), 6 (2024)","journal-title":"Visual Intell."},{"key":"2080_CR39","unstructured":"Dao, T., Gu, A.: Transformers are ssms: generalized models and efficient algorithms through structured state space duality. arXiv preprint arXiv:2405.21060 (2024)"},{"key":"2080_CR40","unstructured":"Wang, X., Wang, S., Ding, Y., Li, Y., Wu, W., Rong, Y., Kong, W., Huang, J., Li, S., Yang, H., et al.: State space model for new-generation network alternative to transformers: a survey. arXiv preprint arXiv:2404.09516 (2024)"},{"key":"2080_CR41","first-page":"103031","volume":"37","author":"Y Liu","year":"2024","unstructured":"Liu, Y., Tian, Y., Zhao, Y., Yu, H., Xie, L., Wang, Y., Ye, Q., Jiao, J., Liu, Y.: Vmamba: visual state space model. Adv. Neural. Inf. Process. Syst. 37, 103031\u2013103063 (2024)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2080_CR42","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"2080_CR43","doi-asserted-by":"crossref","unstructured":"Xiao, M., Zheng, S., Liu, C., Wang, Y., He, D., Ke, G., Bian, J., Lin, Z., Liu, T.-Y.: Invertible image rescaling. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part I 16, pp. 126\u2013144 (2020). Springer","DOI":"10.1007\/978-3-030-58452-8_8"},{"key":"2080_CR44","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Bai, H., Zhang, J., Zhang, Y., Xu, S., Lin, Z., Timofte, R., Van\u00a0Gool, L.: Cddfuse: correlation-driven dual-branch feature decomposition for multi-modality image fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5906\u20135916 (2023)","DOI":"10.1109\/CVPR52729.2023.00572"},{"issue":"10","key":"2080_CR45","doi-asserted-by":"publisher","first-page":"6700","DOI":"10.1109\/TCSVT.2022.3168279","volume":"32","author":"Y Sun","year":"2022","unstructured":"Sun, Y., Cao, B., Zhu, P., Hu, Q.: Drone-based rgb-infrared cross-modality vehicle detection via uncertainty-aware learning. IEEE Trans. Circ. Syst. Video Technol. 32(10), 6700\u20136713 (2022)","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"key":"2080_CR46","doi-asserted-by":"crossref","unstructured":"Zhang, P., Zhao, J., Wang, D., Lu, H., Ruan, X.: Visible-thermal UAV tracking: a large-scale benchmark and new baseline. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8886\u20138895 (2022)","DOI":"10.1109\/CVPR52688.2022.00868"},{"key":"2080_CR47","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","volume":"83","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Zhang, H., Jiang, X., Ma, J.: Piafusion: a progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion 83, 79\u201392 (2022)","journal-title":"Inf. Fusion"},{"key":"2080_CR48","doi-asserted-by":"crossref","unstructured":"Liu, J., Fan, X., Huang, Z., Wu, G., Liu, R., Zhong, W., Luo, Z.: Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5802\u20135811 (2022)","DOI":"10.1109\/CVPR52688.2022.00571"},{"key":"2080_CR49","first-page":"12484","volume":"34","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Le, Z., Jiang, J., Guo, X.: Fusiondn: a unified densely connected network for image fusion. Proc. AAAI Conf. Artif. Intell. 34, 12484\u201312491 (2020)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"2080_CR50","doi-asserted-by":"crossref","unstructured":"Zhang, M.M., Choi, J., Daniilidis, K., Wolf, M.T., Kanan, C.: Vais: a dataset for recognizing maritime imagery in the visible and infrared spectrums. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 10\u201316 (2015)","DOI":"10.1109\/CVPRW.2015.7301291"},{"key":"2080_CR51","doi-asserted-by":"crossref","unstructured":"Zhang, H., Ma, J.: Sdnet: a versatile squeeze-and-decomposition network for real-time image fusion. Int. J. Comput. Vision 1\u201325 (2021)","DOI":"10.1007\/s11263-021-01501-8"},{"issue":"1","key":"2080_CR52","doi-asserted-by":"publisher","first-page":"37","DOI":"10.1007\/s44267-024-00072-9","volume":"2","author":"X Xie","year":"2024","unstructured":"Xie, X., Cui, Y., Tan, T., Zheng, X., Yu, Z.: Fusionmamba: dynamic feature enhancement for multimodal image fusion with mamba. Visual Intell. 2(1), 37 (2024)","journal-title":"Visual Intell."},{"key":"2080_CR53","doi-asserted-by":"crossref","unstructured":"Liu, J., Zhang, B., Mei, Q., Li, X., Zou, Y., Jiang, Z., Ma, L., Liu, R., Fan, X.: Dcevo: discriminative cross-dimensional evolutionary learning for infrared and visible image fusion. arXiv preprint arXiv:2503.17673 (2025)","DOI":"10.1109\/CVPR52734.2025.00213"},{"key":"2080_CR54","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Xu, S., Zhang, C., Liu, J., Zhang, J., Li, P.: Didfuse: deep image decomposition for infrared and visible image fusion. In: IJCAI, pp. 970\u2013976. ijcai.org, Yokohama, Japan (2020)","DOI":"10.24963\/ijcai.2020\/135"},{"key":"2080_CR55","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Bai, H., Zhang, J., Zhang, Y., Zhang, K., Xu, S., Chen, D., Timofte, R., Van\u00a0Gool, L.: Equivariant multi-modality image fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 25912\u201325921 (2024)","DOI":"10.1109\/CVPR52733.2024.02448"},{"key":"2080_CR56","doi-asserted-by":"crossref","unstructured":"Chen, M., Cheng, Y., He, X., Wang, X., Aze, Y., Xiang, J.: Simplefusion: a simple fusion framework for infrared and visible images. In: Chinese Conference on Pattern Recognition and Computer Vision (PRCV), pp. 49\u201363 (2024). Springer","DOI":"10.1007\/978-981-97-8685-5_4"},{"issue":"7","key":"2080_CR57","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma, J., Tang, L., Fan, F., Huang, J., Mei, X., Ma, Y.: Swinfusion: cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA J. Automatica Sinica 9(7), 1200\u20131217 (2022)","journal-title":"IEEE\/CAA J. Automatica Sinica"},{"issue":"5","key":"2080_CR58","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.-J.: Densefuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"2080_CR59","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. (2020)"},{"key":"2080_CR60","unstructured":"Khanam, R., Hussain, M.: Yolov11: an overview of the key architectural enhancements. arXiv preprint arXiv:2410.17725 (2024)"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02080-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-02080-1","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02080-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T04:20:08Z","timestamp":1770783608000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-02080-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,4]]},"references-count":60,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["2080"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-02080-1","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,4]]},"assertion":[{"value":"22 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 November 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"32"}}