{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T02:17:05Z","timestamp":1773195425047,"version":"3.50.1"},"reference-count":87,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2025,3,11]],"date-time":"2025-03-11T00:00:00Z","timestamp":1741651200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,3,11]],"date-time":"2025-03-11T00:00:00Z","timestamp":1741651200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100004826","name":"Natural Science Foundation of Beijing Municipality","doi-asserted-by":"publisher","award":["BJXZ2021-012-00046"],"award-info":[{"award-number":["BJXZ2021-012-00046"]}],"id":[{"id":"10.13039\/501100004826","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004826","name":"Natural Science Foundation of Beijing Municipality","doi-asserted-by":"publisher","award":["BJXZ2021-012-00046"],"award-info":[{"award-number":["BJXZ2021-012-00046"]}],"id":[{"id":"10.13039\/501100004826","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004826","name":"Natural Science Foundation of Beijing Municipality","doi-asserted-by":"publisher","award":["BJXZ2021-012-00046"],"award-info":[{"award-number":["BJXZ2021-012-00046"]}],"id":[{"id":"10.13039\/501100004826","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004826","name":"Natural Science Foundation of Beijing Municipality","doi-asserted-by":"publisher","award":["BJXZ2021-012-00046"],"award-info":[{"award-number":["BJXZ2021-012-00046"]}],"id":[{"id":"10.13039\/501100004826","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004826","name":"Natural Science Foundation of Beijing Municipality","doi-asserted-by":"publisher","award":["BJXZ2021-012-00046"],"award-info":[{"award-number":["BJXZ2021-012-00046"]}],"id":[{"id":"10.13039\/501100004826","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004826","name":"Natural Science Foundation of Beijing Municipality","doi-asserted-by":"publisher","award":["BJXZ2021-012-00046"],"award-info":[{"award-number":["BJXZ2021-012-00046"]}],"id":[{"id":"10.13039\/501100004826","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004826","name":"Natural Science Foundation of Beijing Municipality","doi-asserted-by":"publisher","award":["BJXZ2021-012-00046"],"award-info":[{"award-number":["BJXZ2021-012-00046"]}],"id":[{"id":"10.13039\/501100004826","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s00371-025-03840-w","type":"journal-article","created":{"date-parts":[[2025,3,11]],"date-time":"2025-03-11T11:15:36Z","timestamp":1741691736000},"page":"7817-7838","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Enhancing infrared and visible image fusion through multiscale Gaussian total variation and adaptive local entropy"],"prefix":"10.1007","volume":"41","author":[{"given":"Hao","family":"Li","sequence":"first","affiliation":[]},{"given":"Shengkun","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Deng","sequence":"additional","affiliation":[]},{"given":"Chenhua","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Yifan","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Hanrui","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Heng","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Mingli","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Lianqing","family":"Zhu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,3,11]]},"reference":[{"issue":"2","key":"3840_CR1","doi-asserted-by":"publisher","first-page":"599","DOI":"10.3390\/s23020599","volume":"23","author":"W Ma","year":"2023","unstructured":"Ma, W., Wang, K., Li, J., Yang, S.X., Li, J., Song, L., Li, Q.: Infrared and visible image fusion technology and application: a review. Sensors 23(2), 599 (2023)","journal-title":"Sensors"},{"key":"3840_CR2","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"J Ma","year":"2019","unstructured":"Ma, J., Ma, Y., Li, C.: Infrared and visible image fusion methods and applications: a survey. Inf. Fusion 45, 153\u2013178 (2019)","journal-title":"Inf. Fusion"},{"issue":"5","key":"3840_CR3","doi-asserted-by":"publisher","first-page":"441","DOI":"10.3390\/rs9050441","volume":"9","author":"H Li","year":"2017","unstructured":"Li, H., Ding, W., Cao, X., Liu, C.: Image registration and fusion of visible and infrared integrated camera for medium-altitude unmanned aerial vehicle remote sensing. Remote Sens. 9(5), 441 (2017)","journal-title":"Remote Sens."},{"key":"3840_CR4","doi-asserted-by":"crossref","unstructured":"Yang, Y., Han, C., Kang, X., Han, D.: An overview on pixel-level image fusion in remote sensing. In: 2007 IEEE International Conference on Automation and Logistics, pp. 2339\u20132344. IEEE (2007)","DOI":"10.1109\/ICAL.2007.4338968"},{"issue":"2","key":"3840_CR5","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1080\/17538947.2013.869266","volume":"7","author":"C Pohl","year":"2014","unstructured":"Pohl, C., Genderen, J.: Remote sensing image fusion: an update in the context of digital earth. Int. J. Digi. Earth 7(2), 158\u2013172 (2014)","journal-title":"Int. J. Digi. Earth"},{"key":"3840_CR6","doi-asserted-by":"publisher","first-page":"590","DOI":"10.1007\/s11432-011-4536-9","volume":"55","author":"H Liu","year":"2012","unstructured":"Liu, H., Sun, F.: Fusion tracking in color and infrared images using joint sparse representation. Sci. China Inf. Sci. 55, 590\u2013599 (2012)","journal-title":"Sci. China Inf. Sci."},{"issue":"12","key":"3840_CR7","doi-asserted-by":"publisher","first-page":"1696","DOI":"10.1109\/TKDE.2006.183","volume":"18","author":"D Smith","year":"2006","unstructured":"Smith, D., Singh, S.: Approaches to multisensor data fusion in target tracking: A survey. IEEE Trans. Knowl. Data Eng. 18(12), 1696\u20131710 (2006)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"issue":"5","key":"3840_CR8","doi-asserted-by":"publisher","first-page":"7757","DOI":"10.1007\/s11042-022-13569-6","volume":"82","author":"A Khmag","year":"2023","unstructured":"Khmag, A.: Additive gaussian noise removal based on generative adversarial network model and semi-soft thresholding approach. Multimed. Tools Appl. 82(5), 7757\u20137777 (2023)","journal-title":"Multimed. Tools Appl."},{"issue":"21","key":"3840_CR9","doi-asserted-by":"publisher","first-page":"15523","DOI":"10.1007\/s00500-023-09148-y","volume":"27","author":"A Khmag","year":"2023","unstructured":"Khmag, A.: Natural digital image mixed noise removal using regularization Perona\u2013Malik model and pulse coupled neural networks. Soft Comput. 27(21), 15523\u201315532 (2023)","journal-title":"Soft Comput."},{"issue":"23","key":"3840_CR10","doi-asserted-by":"publisher","first-page":"17505","DOI":"10.1007\/s00500-023-09204-7","volume":"27","author":"A Khmag","year":"2023","unstructured":"Khmag, A.: Smoke removal technique of industrial scene images based on second-generation wavelets and dark channel prior model. Soft Comput. 27(23), 17505\u201317514 (2023)","journal-title":"Soft Comput."},{"key":"3840_CR11","doi-asserted-by":"crossref","unstructured":"Conaire, C.O., O\u2019Connor, N.E., Cooke, E., Smeaton, A.F.: Comparison of fusion methods for thermo-visual surveillance tracking. In: 2006 9th International Conference on Information Fusion, pp. 1\u20137. IEEE (2006)","DOI":"10.1109\/ICIF.2006.301618"},{"key":"3840_CR12","doi-asserted-by":"crossref","unstructured":"Kumar, P., Mittal, A., Kumar, P.: Fusion of thermal infrared and visible spectrum video for robust surveillance. In: Computer Vision, Graphics and Image Processing: 5th Indian Conference, ICVGIP 2006, Madurai, India, December 13-16, 2006. Proceedings, pp. 528\u2013539. Springer (2006)","DOI":"10.1007\/11949619_47"},{"issue":"6","key":"3840_CR13","doi-asserted-by":"publisher","first-page":"1771","DOI":"10.1016\/j.patcog.2006.11.010","volume":"40","author":"J Han","year":"2007","unstructured":"Han, J., Bhanu, B.: Fusion of color and infrared video for moving human detection. Pattern Recognit. 40(6), 1771\u20131784 (2007)","journal-title":"Pattern Recognit."},{"issue":"1","key":"3840_CR14","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1049\/iet-ipr.2009.0374","volume":"5","author":"I Ulusoy","year":"2011","unstructured":"Ulusoy, I., Yuruk, H.: New method for the fusion of complementary information from infrared and visual images for object detection. IET Image Process. 5(1), 36\u201348 (2011)","journal-title":"IET Image Process."},{"issue":"11","key":"3840_CR15","doi-asserted-by":"publisher","first-page":"1981","DOI":"10.1364\/OL.38.001981","volume":"38","author":"S Gao","year":"2013","unstructured":"Gao, S., Cheng, Y., Zhao, Y.: Method of visual and infrared fusion for moving object detection. Opt. Lett. 38(11), 1981\u20131983 (2013)","journal-title":"Opt. Lett."},{"key":"3840_CR16","first-page":"129","volume":"4","author":"AA Pure","year":"2013","unstructured":"Pure, A.A., Gupta, N., Shrivastava, M.: An overview of different image fusion methods for medical applications. Int. J. Sci. Eng. Res. 4, 129\u2013133 (2013)","journal-title":"Int. J. Sci. Eng. Res."},{"key":"3840_CR17","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1016\/j.neucom.2015.07.160","volume":"215","author":"J Du","year":"2016","unstructured":"Du, J., Li, W., Lu, K., Xiao, B.: An overview of multi-modal medical image fusion. Neurocomputing 215, 3\u201320 (2016)","journal-title":"Neurocomputing"},{"key":"3840_CR18","doi-asserted-by":"publisher","first-page":"8279342","DOI":"10.1155\/2020\/8279342","volume":"2020","author":"B Huang","year":"2020","unstructured":"Huang, B., Yang, F., Yin, M., Mo, X., Zhong, C., et al.: A review of multimodal medical image fusion techniques. Comput. Math. Methods Med. 2020, 8279342 (2020)","journal-title":"Comput. Math. Methods Med."},{"key":"3840_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.sigpro.2021.108036","volume":"183","author":"H Hermessi","year":"2021","unstructured":"Hermessi, H., Mourali, O., Zagrouba, E.: Multimodal medical image fusion review: theoretical background and recent advances. Signal Process. 183, 108036 (2021)","journal-title":"Signal Process."},{"issue":"1","key":"3840_CR20","first-page":"1188974","volume":"2022","author":"C Liu","year":"2022","unstructured":"Liu, C., Chen, S., Huang, J.: Machine vision-based object detection strategy for weld area. Sci. Program. 2022(1), 1188974 (2022)","journal-title":"Sci. Program."},{"issue":"4","key":"3840_CR21","doi-asserted-by":"publisher","first-page":"338","DOI":"10.1016\/j.vrih.2022.08.017","volume":"5","author":"X Liu","year":"2023","unstructured":"Liu, X., Wu, Z., Wang, X.: The validity analysis of the non-local mean filter and a derived novel denoising method. Virtual Real. Intell. Hardw. 5(4), 338\u2013350 (2023)","journal-title":"Virtual Real. Intell. Hardw."},{"issue":"4","key":"3840_CR22","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.vrih.2023.06.011","volume":"6","author":"X Senhua","year":"2024","unstructured":"Senhua, X., Liqing, G., Liang, W., Wei, F.: Multi-scale context-aware network for continuous sign language recognition. Virtual Real. Intell. Hardw. 6(4), 323\u2013337 (2024)","journal-title":"Virtual Real. Intell. Hardw."},{"key":"3840_CR23","doi-asserted-by":"publisher","first-page":"6821","DOI":"10.1109\/TMM.2022.3214776","volume":"25","author":"J Zhu","year":"2022","unstructured":"Zhu, J., Zhang, Q., Fei, L., Cai, R., Xie, Y., Sheng, B., Yang, X.: FFFN: Frame-by-frame feedback fusion network for video super-resolution. IEEE Trans. Multimed. 25, 6821\u20136835 (2022)","journal-title":"IEEE Trans. Multimed."},{"key":"3840_CR24","doi-asserted-by":"publisher","first-page":"880","DOI":"10.1109\/TIP.2021.3136619","volume":"31","author":"A Nazir","year":"2021","unstructured":"Nazir, A., Cheema, M.N., Sheng, B., Li, P., Li, H., Xue, G., Qin, J., Kim, J., Feng, D.D.: Ecsu-net: an embedded clustering sliced u-net coupled with fusing strategy for efficient intervertebral disc segmentation and classification. IEEE Trans. Image Process. 31, 880\u2013893 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"3840_CR25","doi-asserted-by":"publisher","first-page":"7192","DOI":"10.1109\/TIP.2020.2999854","volume":"29","author":"A Nazir","year":"2020","unstructured":"Nazir, A., Cheema, M.N., Sheng, B., Li, H., Li, P., Yang, P., Jung, Y., Qin, J., Kim, J., Feng, D.D.: OFF-Enet: an optimally fused fully end-to-end network for automatic dense volumetric 3D intracranial blood vessels segmentation. IEEE Trans. Image Process. 29, 7192\u20137202 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"3840_CR26","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-024-03712-9","author":"K Wang","year":"2024","unstructured":"Wang, K., Liu, C., Zhang, R.: CMA-SOD: cross-modal attention fusion network for RGB-D salient object detection. Visual Comput. (2024). https:\/\/doi.org\/10.1007\/s00371-024-03712-9","journal-title":"Visual Comput."},{"issue":"4","key":"3840_CR27","doi-asserted-by":"publisher","first-page":"1005","DOI":"10.1002\/col.22777","volume":"47","author":"B Subramani","year":"2022","unstructured":"Subramani, B., Veluchamy, M.: Cuckoo search optimization-based image color and detail enhancement for contrast distorted images. Color Res. Appl. 47(4), 1005\u20131022 (2022)","journal-title":"Color Res. Appl."},{"issue":"27","key":"3840_CR28","doi-asserted-by":"publisher","first-page":"19945","DOI":"10.1007\/s11042-020-08870-1","volume":"79","author":"M Veluchamy","year":"2020","unstructured":"Veluchamy, M., Subramani, B.: Fuzzy dissimilarity contextual intensity transformation with gamma correction for color image enhancement. Multimed. Tools Appl. 79(27), 19945\u201319961 (2020)","journal-title":"Multimed. Tools Appl."},{"issue":"3","key":"3840_CR29","doi-asserted-by":"publisher","first-page":"3627","DOI":"10.1007\/s11042-022-13409-7","volume":"82","author":"M Veluchamy","year":"2023","unstructured":"Veluchamy, M., Subramani, B.: Artificial bee colony optimized image enhancement framework for invisible images. Multimed. Tools Appl. 82(3), 3627\u20133646 (2023)","journal-title":"Multimed. Tools Appl."},{"issue":"5","key":"3840_CR30","doi-asserted-by":"publisher","first-page":"2293","DOI":"10.1002\/cav.2293","volume":"35","author":"Z Xiao","year":"2024","unstructured":"Xiao, Z., Chen, Y., Zhou, X., He, M., Liu, L., Yu, F., Jiang, M.: Human action recognition in immersive virtual reality based on multi-scale spatio-temporal attention network. Comput. Anim. Virtual Worlds 35(5), 2293 (2024)","journal-title":"Comput. Anim. Virtual Worlds"},{"issue":"3","key":"3840_CR31","doi-asserted-by":"publisher","first-page":"2268","DOI":"10.1002\/cav.2268","volume":"35","author":"Y Bao","year":"2024","unstructured":"Bao, Y., Liu, X., Qi, Y., Liu, R., Li, H.: Adaptive information fusion network for multi-modal personality recognition. Comput. Anim. Virtual Worlds 35(3), 2268 (2024)","journal-title":"Comput. Anim. Virtual Worlds"},{"key":"3840_CR32","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2024.3370752","author":"H Chen","year":"2024","unstructured":"Chen, H., Deng, L., Chen, Z., Liu, C., Zhu, L., Dong, M., Lu, X., Guo, C.: SFCFusion: Spatial-frequency collaborative infrared and visible image fusion. IEEE Trans. Instrum. Meas. (2024). https:\/\/doi.org\/10.1109\/TIM.2024.3370752","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"3840_CR33","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-024-03571-4","author":"P Wu","year":"2024","unstructured":"Wu, P., Tang, J.: FHFN: content and context feature hierarchical fusion networks for multi-focus image fusion. Visual Comput. (2024). https:\/\/doi.org\/10.1007\/s00371-024-03571-4","journal-title":"Visual Comput."},{"key":"3840_CR34","doi-asserted-by":"crossref","unstructured":"Ren, X., Meng, F., Hu, T., Liu, Z., Wang, C.: Infrared-visible image fusion based on convolutional neural networks (CNN). In: Intelligence Science and Big Data Engineering: 8th International Conference, IScIDE 2018, Lanzhou, China, August 18\u201319, 2018, Revised Selected Papers 8, pp. 301\u2013307. Springer (2018)","DOI":"10.1007\/978-3-030-02698-1_26"},{"issue":"14","key":"3840_CR35","doi-asserted-by":"publisher","first-page":"3562","DOI":"10.1049\/iet-ipr.2020.0360","volume":"14","author":"L Li","year":"2020","unstructured":"Li, L., Xia, Z., Han, H., He, G., Roli, F., Feng, X.: Infrared and visible image fusion using a shallow CNN and structural similarity constraint. IET Image Process. 14(14), 3562\u20133571 (2020)","journal-title":"IET Image Process."},{"key":"3840_CR36","doi-asserted-by":"crossref","unstructured":"Li, H., Wu, X.-J., Kittler, J.: Infrared and visible image fusion using a deep learning framework. In: 2018 24th International Conference on Pattern Recognition (ICPR), pp. 2705\u20132710. IEEE (2018)","DOI":"10.1109\/ICPR.2018.8546006"},{"issue":"5","key":"3840_CR37","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.-J.: Densefuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"3840_CR38","doi-asserted-by":"publisher","first-page":"78956","DOI":"10.1109\/ACCESS.2023.3298437","volume":"11","author":"H Wang","year":"2023","unstructured":"Wang, H., Li, L., Li, C., Lu, X.: Infrared and visible image fusion based on autoencoder composed of CNN-transformer. IEEE Access. 11, 78956\u201378969 (2023). https:\/\/doi.org\/10.1109\/ACCESS.2023.3298437","journal-title":"IEEE Access."},{"key":"3840_CR39","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2021.103839","volume":"117","author":"L Ren","year":"2021","unstructured":"Ren, L., Pan, Z., Cao, J., Liao, J.: Infrared and visible image fusion based on variational auto-encoder and infrared feature compensation. Infrared Phys. Technol. 117, 103839 (2021)","journal-title":"Infrared Phys. Technol."},{"key":"3840_CR40","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: Fusiongan: a generative adversarial network for infrared and visible image fusion. Inf. Fusion 48, 11\u201326 (2019)","journal-title":"Inf. Fusion"},{"key":"3840_CR41","doi-asserted-by":"publisher","first-page":"336","DOI":"10.1016\/j.inffus.2022.12.007","volume":"92","author":"Y Rao","year":"2023","unstructured":"Rao, Y., Wu, D., Han, M., Wang, T., Yang, Y., Lei, T., Zhou, C., Bai, H., Xing, L.: AT-GAN: a generative adversarial network with attention and transition for infrared and visible image fusion. Inf. Fusion 92, 336\u2013349 (2023)","journal-title":"Inf. Fusion"},{"key":"3840_CR42","doi-asserted-by":"publisher","first-page":"1383","DOI":"10.1109\/TMM.2020.2997127","volume":"23","author":"J Li","year":"2020","unstructured":"Li, J., Huo, H., Li, C., Wang, R., Feng, Q.: AttentionFGAN: infrared and visible image fusion using attention-based generative adversarial networks. IEEE Trans. Multimed. 23, 1383\u20131396 (2020)","journal-title":"IEEE Trans. Multimed."},{"key":"3840_CR43","doi-asserted-by":"publisher","unstructured":"Tang, W., He, F., Liu, Y.: YDTR: Infrared and visible image fusion via y-shape dynamic transformer. IEEE Trans. Multimed. 25, 5413\u20135428 (2023). https:\/\/doi.org\/10.1109\/TMM.2022.3192661","DOI":"10.1109\/TMM.2022.3192661"},{"key":"3840_CR44","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3218574","volume":"71","author":"J Li","year":"2022","unstructured":"Li, J., Zhu, J., Li, C., Chen, X., Yang, B.: CGTF: convolution-guided transformer for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 71, 1\u201314 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"3840_CR45","doi-asserted-by":"publisher","unstructured":"Yue, J., Fang, L., Xia, S., Deng, Y., Ma, J.: Dif-fusion: toward high color fidelity in infrared and visible image fusion with diffusion models. IEEE Trans. Image Process. 32, 5705\u20135720 (2023). https:\/\/doi.org\/10.1109\/TIP.2023.3322046","DOI":"10.1109\/TIP.2023.3322046"},{"key":"3840_CR46","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Bai, H., Zhu, Y., Zhang, J., Xu, S., Zhang, Y., Zhang, K., Meng, D., Timofte, R., Van\u00a0Gool, L.: DDFM: denoising diffusion model for multi-modality image fusion. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8082\u20138093. (2023)","DOI":"10.1109\/ICCV51070.2023.00742"},{"key":"3840_CR47","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.05.004","volume":"33","author":"S Li","year":"2017","unstructured":"Li, S., Kang, X., Fang, L., Hu, J., Yin, H.: Pixel-level image fusion. A survey of the state of the art. Inf. Fusion 33, 100\u2013112 (2017)","journal-title":"Inf. Fusion"},{"key":"3840_CR48","doi-asserted-by":"publisher","first-page":"16040","DOI":"10.1109\/ACCESS.2017.2735865","volume":"5","author":"A Dogra","year":"2017","unstructured":"Dogra, A., Goyal, B., Agrawal, S.: From multi-scale decomposition to non-multi-scale decomposition methods: a comprehensive survey of image fusion techniques and its applications. IEEE Access 5, 16040\u201316067 (2017)","journal-title":"IEEE Access"},{"key":"3840_CR49","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2024.105351","volume":"140","author":"C Liu","year":"2024","unstructured":"Liu, C., Chen, H., Deng, L., Guo, C., Lu, X., Yu, H., Zhu, L., Dong, M.: Modality specific infrared and visible image fusion based on multi-scale rich feature representation under low-light environment. Infrared Phys. Technol. 140, 105351 (2024)","journal-title":"Infrared Phys. Technol."},{"key":"3840_CR50","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2023.104956","volume":"135","author":"C Guo","year":"2023","unstructured":"Guo, C., Liu, C., Deng, L., Chen, Z., Dong, M., Zhu, L., Chen, H., Lu, X.: Multi-scale infrared and visible image fusion framework based on dual partial differential equations. Infrared Phys. Technol. 135, 104956 (2023)","journal-title":"Infrared Phys. Technol."},{"key":"3840_CR51","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2020.2986875","volume":"70","author":"Y Yang","year":"2020","unstructured":"Yang, Y., Zhang, Y., Huang, S., Zuo, Y., Sun, J.: Infrared and visible image fusion using visual saliency sparse representation and detail injection model. IEEE Trans. Instrum. Meas. 70, 1\u201315 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"3840_CR52","doi-asserted-by":"publisher","first-page":"286","DOI":"10.1016\/j.infrared.2015.10.004","volume":"73","author":"B Zhang","year":"2015","unstructured":"Zhang, B., Lu, X., Pei, H., Zhao, Y.: A fusion algorithm for infrared and visible images based on saliency analysis and non-subsampled Shearlet transform. Infrared Phys. Technol. 73, 286\u2013297 (2015)","journal-title":"Infrared Phys. Technol."},{"key":"3840_CR53","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1016\/j.inffus.2015.11.003","volume":"30","author":"Z Zhou","year":"2016","unstructured":"Zhou, Z., Wang, B., Li, S., Dong, M.: Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters. Inf. Fusion 30, 15\u201326 (2016)","journal-title":"Inf. Fusion"},{"key":"3840_CR54","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijleo.2022.168914","volume":"258","author":"Y Luo","year":"2022","unstructured":"Luo, Y., He, K., Xu, D., Yin, W., Liu, W.: Infrared and visible image fusion based on visibility enhancement and hybrid multiscale decomposition. Optik 258, 168914 (2022)","journal-title":"Optik"},{"issue":"6","key":"3840_CR55","first-page":"1","volume":"31","author":"L Xu","year":"2012","unstructured":"Xu, L., Yan, Q., Xia, Y., Jia, J.: Structure extraction from texture via relative total variation. ACM Trans. Gr. (TOG) 31(6), 1\u201310 (2012)","journal-title":"ACM Trans. Gr. (TOG)"},{"key":"3840_CR56","doi-asserted-by":"publisher","first-page":"12","DOI":"10.1016\/j.neucom.2016.03.009","volume":"202","author":"Y Ma","year":"2016","unstructured":"Ma, Y., Chen, J., Chen, C., Fan, F., Ma, J.: Infrared and visible image fusion using total variation model. Neurocomputing 202, 12\u201319 (2016)","journal-title":"Neurocomputing"},{"key":"3840_CR57","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.02.001","volume":"31","author":"J Ma","year":"2016","unstructured":"Ma, J., Chen, C., Li, C., Huang, J.: Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fusion 31, 100\u2013109 (2016)","journal-title":"Inf. Fusion"},{"key":"3840_CR58","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2022.104112","volume":"123","author":"J Chen","year":"2022","unstructured":"Chen, J., Li, X., Wu, K.: Infrared and visible image fusion based on relative total variation decomposition. Infrared Phys. Technol. 123, 104112 (2022)","journal-title":"Infrared Phys. Technol."},{"key":"3840_CR59","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen, J., Li, X., Luo, L., Mei, X., Ma, J.: Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 508, 64\u201378 (2020)","journal-title":"Inf. Sci."},{"key":"3840_CR60","doi-asserted-by":"publisher","first-page":"154","DOI":"10.1016\/j.infrared.2018.06.002","volume":"92","author":"T Ma","year":"2018","unstructured":"Ma, T., Ma, J., Fang, B., Hu, F., Quan, S., Du, H.: Multi-scale decomposition based fusion of infrared and visible image via total variation and saliency analysis. Infrared Phys. Technol. 92, 154\u2013162 (2018)","journal-title":"Infrared Phys. Technol."},{"key":"3840_CR61","doi-asserted-by":"crossref","unstructured":"Burt, P.J., Adelson, E.H.: The Laplacian pyramid as a compact image code. In: Readings in Computer Vision, pp. 671\u2013679. Elsevier (1987)","DOI":"10.1016\/B978-0-08-051581-6.50065-9"},{"key":"3840_CR62","doi-asserted-by":"crossref","unstructured":"Chipman, L.J., Orr, T.M., Graham, L.N.: Wavelets and image fusion. In: Proceedings., International Conference on Image Processing, vol. 3, pp. 248\u2013251. IEEE (1995)","DOI":"10.1109\/ICIP.1995.537627"},{"issue":"12","key":"3840_CR63","doi-asserted-by":"publisher","first-page":"2091","DOI":"10.1109\/TIP.2005.859376","volume":"14","author":"MN Do","year":"2005","unstructured":"Do, M.N., Vetterli, M.: The contourlet transform: an efficient directional multiresolution image representation. IEEE Trans. Image Process. 14(12), 2091\u20132106 (2005)","journal-title":"IEEE Trans. Image Process."},{"issue":"10","key":"3840_CR64","doi-asserted-by":"publisher","first-page":"3089","DOI":"10.1109\/TIP.2006.877507","volume":"15","author":"AL Da Cunha","year":"2006","unstructured":"Da Cunha, A.L., Zhou, J., Do, M.N.: The nonsubsampled contourlet transform: theory, design, and applications. IEEE Trans. Image Process. 15(10), 3089\u20133101 (2006)","journal-title":"IEEE Trans. Image Process."},{"key":"3840_CR65","doi-asserted-by":"crossref","unstructured":"Bavirisetti, D.P., Xiao, G., Liu, G.: Multi-sensor image fusion based on fourth order partial differential equations. In: 2017 20th International Conference on Information Fusion (Fusion), pp. 1\u20139. IEEE (2017)","DOI":"10.23919\/ICIF.2017.8009719"},{"key":"3840_CR66","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","volume":"82","author":"J Ma","year":"2017","unstructured":"Ma, J., Zhou, Z., Wang, B., Zong, H.: Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 82, 8\u201317 (2017)","journal-title":"Infrared Phys. Technol."},{"key":"3840_CR67","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.128116","volume":"600","author":"G Yang","year":"2024","unstructured":"Yang, G., Li, J., Lei, H., Gao, X.: A multi-scale information integration framework for infrared and visible image fusion. Neurocomputing 600, 128116 (2024)","journal-title":"Neurocomputing"},{"key":"3840_CR68","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2024.107905","volume":"132","author":"Y Zhou","year":"2024","unstructured":"Zhou, Y., He, K., Xu, D., Tao, D., Lin, X., Li, C.: ASFusion: Adaptive visual enhancement and structural patch decomposition for infrared and visible image fusion. Eng. Appl. Artif. Intell. 132, 107905 (2024)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"3840_CR69","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1016\/j.inffus.2014.09.004","volume":"24","author":"Y Liu","year":"2015","unstructured":"Liu, Y., Liu, S., Wang, Z.: A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 24, 147\u2013164 (2015)","journal-title":"Inf. Fusion"},{"issue":"17","key":"3840_CR70","doi-asserted-by":"publisher","first-page":"4980","DOI":"10.1016\/j.ijleo.2014.04.006","volume":"125","author":"Y Chen","year":"2014","unstructured":"Chen, Y., Xiong, J., Liu, H.-L., Fan, Q.: Fusion method of infrared and visible images based on neighborhood characteristic and regionalization in NSCT domain. Optik 125(17), 4980\u20134984 (2014)","journal-title":"Optik"},{"key":"3840_CR71","doi-asserted-by":"publisher","DOI":"10.1016\/j.optlaseng.2022.107268","volume":"160","author":"D Zou","year":"2023","unstructured":"Zou, D., Yang, B.: Infrared and low-light visible image fusion based on hybrid multiscale decomposition and adaptive light adjustment. Opt. Lasers Eng. 160, 107268 (2023)","journal-title":"Opt. Lasers Eng."},{"issue":"2","key":"3840_CR72","first-page":"57","volume":"28","author":"L Zhan","year":"2017","unstructured":"Zhan, L., Zhuang, Y., Huang, L.: Infrared and visible images fusion method based on discrete wavelet transform. J. Comput. 28(2), 57\u201371 (2017)","journal-title":"J. Comput."},{"key":"3840_CR73","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2023.104767","volume":"136","author":"L Dong","year":"2024","unstructured":"Dong, L., Wang, J.: Infrared and visible light image fusion via pixel mean shift and source image gradient. Infrared Phys. Technol. 136, 104767 (2024)","journal-title":"Infrared Phys. Technol."},{"issue":"1\u20132","key":"3840_CR74","doi-asserted-by":"publisher","first-page":"75","DOI":"10.1016\/j.infrared.2004.03.011","volume":"46","author":"C Ibarra-Castanedo","year":"2004","unstructured":"Ibarra-Castanedo, C., Gonzalez, D., Klein, M., Pilla, M., Vallerand, S., Maldague, X.: Infrared image processing and data analysis. Infrared Phys. Technol. 46(1\u20132), 75\u201383 (2004)","journal-title":"Infrared Phys. Technol."},{"issue":"4","key":"3840_CR75","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"3840_CR76","doi-asserted-by":"publisher","DOI":"10.1117\/1.2945910","volume":"2","author":"JW Roberts","year":"2008","unstructured":"Roberts, J.W., Van Aardt, J.A., Ahmed, F.B.: Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2(1), 023522 (2008)","journal-title":"J. Appl. Remote Sens."},{"key":"3840_CR77","doi-asserted-by":"publisher","first-page":"338","DOI":"10.1007\/s10278-007-9044-5","volume":"21","author":"D-Y Tsai","year":"2008","unstructured":"Tsai, D.-Y., Lee, Y., Matsuyama, E.: Information entropy measure for evaluation of image quality. J. Digit. Imaging 21, 338\u2013347 (2008)","journal-title":"J. Digit. Imaging"},{"key":"3840_CR78","doi-asserted-by":"crossref","unstructured":"Zhang, X., Ye, P., Xiao, G.: Vifb: A visible and infrared image fusion benchmark. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 104\u2013105 (2020)","DOI":"10.1109\/CVPRW50498.2020.00060"},{"key":"3840_CR79","doi-asserted-by":"crossref","unstructured":"Liu, J., Fan, X., Huang, Z., Wu, G., Liu, R., Zhong, W., Luo, Z.: Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5802\u20135811 (2022)","DOI":"10.1109\/CVPR52688.2022.00571"},{"issue":"1","key":"3840_CR80","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1109\/JSEN.2015.2478655","volume":"16","author":"DP Bavirisetti","year":"2015","unstructured":"Bavirisetti, D.P., Dhuli, R.: Fusion of infrared and visible sensor images based on anisotropic diffusion and Karhunen-Loeve transform. IEEE Sens. J. 16(1), 203\u2013209 (2015)","journal-title":"IEEE Sens. J."},{"key":"3840_CR81","doi-asserted-by":"publisher","first-page":"1193","DOI":"10.1007\/s11760-013-0556-9","volume":"9","author":"B Shreyamsha Kumar","year":"2015","unstructured":"Shreyamsha Kumar, B.: Image fusion based on pixel significance using cross bilateral filter. Signal, Image Video Process. 9, 1193\u20131204 (2015)","journal-title":"Signal, Image Video Process."},{"key":"3840_CR82","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1016\/j.inffus.2015.11.003","volume":"30","author":"Z Zhou","year":"2016","unstructured":"Zhou, Z., Wang, B., Li, S., Dong, M.: Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with gaussian and bilateral filters. Inf. Fusion 30, 15\u201326 (2016)","journal-title":"Inf. Fusion"},{"key":"3840_CR83","doi-asserted-by":"publisher","first-page":"5576","DOI":"10.1007\/s00034-019-01131-z","volume":"38","author":"DP Bavirisetti","year":"2019","unstructured":"Bavirisetti, D.P., Xiao, G., Zhao, J., Dhuli, R., Liu, G.: Multi-scale guided image and video fusion: a fast and efficient approach. Circuits Syst. Signal Process. 38, 5576\u20135605 (2019)","journal-title":"Circuits Syst. Signal Process."},{"key":"3840_CR84","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3216413","volume":"71","author":"Z Wang","year":"2022","unstructured":"Wang, Z., Wu, Y., Wang, J., Xu, J., Shao, W.: Res2fusion: infrared and visible image fusion based on dense res2net and double nonlocal attention models. IEEE Trans. Instrum. Meas. 71, 1\u201312 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"3840_CR85","doi-asserted-by":"crossref","unstructured":"Tang, W., He, F., Liu, Y., Duan, Y., Si, T.: DATFuse: infrared and visible image fusion via dual attention transformer. IEEE Trans. Circuits Syst. Video Technol. 33(7), 3159\u20133172. https:\/\/doi.org\/10.1109\/TCSVT.2023.3234340 (2023)","DOI":"10.1109\/TCSVT.2023.3234340"},{"key":"3840_CR86","doi-asserted-by":"crossref","unstructured":"Wang, Z., Wang, J., Wu, Y., Xu, J., Zhang, X.: UNFusion: a unified multi-scale densely connected network for infrared and visible image fusion. IEEE Trans. Circuits Syst. Video Technol. 32(6):3360\u20133374 (2021)","DOI":"10.1109\/TCSVT.2021.3109895"},{"key":"3840_CR87","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2023.110192","volume":"148","author":"X Luo","year":"2024","unstructured":"Luo, X., Wang, J., Zhang, Z., Wu, X.-J.: A full-scale hierarchical encoder-decoder network with cascading edge-prior for infrared and visible image fusion. Pattern Recognit. 148, 110192 (2024)","journal-title":"Pattern Recognit."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-03840-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-03840-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-03840-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T09:47:25Z","timestamp":1757152045000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-03840-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,3,11]]},"references-count":87,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["3840"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-03840-w","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,3,11]]},"assertion":[{"value":"5 February 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 March 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there are no potential or explicit financial interests in this paper and express no objections to the order of authorship.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}