{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,19]],"date-time":"2025-11-19T15:49:32Z","timestamp":1763567372853,"version":"3.45.0"},"reference-count":65,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2025,10,31]],"date-time":"2025-10-31T00:00:00Z","timestamp":1761868800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,10,31]],"date-time":"2025-10-31T00:00:00Z","timestamp":1761868800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61806220","62273356"],"award-info":[{"award-number":["61806220","62273356"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J. King Saud Univ. Comput. Inf. Sci."],"published-print":{"date-parts":[[2025,11]]},"DOI":"10.1007\/s44443-025-00309-7","type":"journal-article","created":{"date-parts":[[2025,10,31]],"date-time":"2025-10-31T15:53:29Z","timestamp":1761926009000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["FIVFusion: Fog-free infrared and visible image fusion"],"prefix":"10.1007","volume":"37","author":[{"given":"Xianghe","family":"Bi","sequence":"first","affiliation":[]},{"given":"Yang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Wenyu","family":"Ji","sequence":"additional","affiliation":[]},{"given":"Jiabao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Ruizhi","family":"Fu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2289-4589","authenticated-orcid":false,"given":"Zhuang","family":"Miao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,31]]},"reference":[{"key":"309_CR1","doi-asserted-by":"publisher","unstructured":"Bavirisetti DP, Xiao G, Liu G (2017) Multi-sensor image fusion based on fourth order partial differential equations. In: 2017 20th International conference on information fusion (Fusion). https:\/\/doi.org\/10.23919\/icif.2017.8009719","DOI":"10.23919\/icif.2017.8009719"},{"key":"309_CR2","doi-asserted-by":"publisher","unstructured":"Bavirisetti DP, Xiao G, Zhao J, Dhuli R, Liu G (2019) Multi-scale guided image and video fusion: A fast and efficient approach. Circuits, Syst, Signal Process 5576\u20135605. https:\/\/doi.org\/10.1007\/s00034-019-01131-z","DOI":"10.1007\/s00034-019-01131-z"},{"issue":"1","key":"309_CR3","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1109\/JSEN.2015.2478655","volume":"16","author":"DP Bavirisetti","year":"2015","unstructured":"Bavirisetti DP, Dhuli R (2015) Fusion of infrared and visible sensor images based on anisotropic diffusion and karhunen-loeve transform. IEEE Sens J 16(1):203\u2013209","journal-title":"IEEE Sens J"},{"issue":"3","key":"309_CR4","doi-asserted-by":"publisher","first-page":"551","DOI":"10.1007\/s10772-021-09822-2","volume":"25","author":"D Bhavana","year":"2022","unstructured":"Bhavana D, Kishore Kumar K, Ravi Tej D (2022) Infrared and visible image fusion using latent low rank technique for surveillance applications. Int J Speech Technol 25(3):551\u2013560","journal-title":"Int J Speech Technol"},{"issue":"11","key":"309_CR5","doi-asserted-by":"publisher","first-page":"5187","DOI":"10.1109\/TIP.2016.2598681","volume":"25","author":"B Cai","year":"2016","unstructured":"Cai B, Xu X, Jia K, Qing C, Tao D (2016) Dehazenet: an end-to-end system for single image haze removal. IEEE Trans Image Process 25(11):5187\u20135198","journal-title":"IEEE Trans Image Process"},{"key":"309_CR6","doi-asserted-by":"publisher","first-page":"2077","DOI":"10.1109\/TIP.2023.3263113","volume":"32","author":"Z Chang","year":"2023","unstructured":"Chang Z, Feng Z, Yang S, Gao Q (2023) Aft: adaptive fusion transformer for visible and infrared images. IEEE Trans Image Process 32:2077\u20132092","journal-title":"IEEE Trans Image Process"},{"issue":"10","key":"309_CR7","doi-asserted-by":"publisher","first-page":"1421","DOI":"10.1016\/j.imavis.2007.12.002","volume":"27","author":"Y Chen","year":"2009","unstructured":"Chen Y, Blum RS (2009) A new automated quality assessment algorithm for image fusion. Image Vis Comput 27(10):1421\u20131432","journal-title":"Image Vis Comput"},{"key":"309_CR8","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen J, Li X, Luo L, Mei X, Ma J (2020) Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf Sci 508:64\u201378","journal-title":"Inf Sci"},{"key":"309_CR9","doi-asserted-by":"publisher","first-page":"655","DOI":"10.1109\/TMM.2021.3057493","volume":"24","author":"J Chen","year":"2021","unstructured":"Chen J, Li X, Luo L, Ma J (2021) Multi-focus image fusion based on multi-scale gradients and image matting. IEEE Trans Multimedia 24:655\u2013667","journal-title":"IEEE Trans Multimedia"},{"key":"309_CR10","doi-asserted-by":"crossref","unstructured":"Cui G, Feng H, Xu Z, Li Q, Chen Y (2015) Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Optics Commun 341:199\u2013209","DOI":"10.1016\/j.optcom.2014.12.032"},{"key":"309_CR11","doi-asserted-by":"crossref","unstructured":"Dong J, Pan J (2020) Physics-based feature dehazing networks. In: Computer Vision\u2013ECCV 2020: 16th European conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXX 16. Springer, pp 188\u2013204","DOI":"10.1007\/978-3-030-58577-8_12"},{"key":"309_CR12","unstructured":"Dosovitskiy A (2020) An image is worth 16x16 words: transformers for image recognition at scale. arXiv:2010.11929"},{"issue":"12","key":"309_CR13","doi-asserted-by":"publisher","first-page":"2959","DOI":"10.1109\/26.477498","volume":"43","author":"AM Eskicioglu","year":"1995","unstructured":"Eskicioglu AM, Fisher PS (1995) Image quality measures and their performance. IEEE Trans Commun 43(12):2959\u20132965","journal-title":"IEEE Trans Commun"},{"key":"309_CR14","doi-asserted-by":"crossref","unstructured":"Feng C, Zhuo S, Zhang X, Shen L, S\u00fcsstrunk S (2013) Near-infrared guided color image dehazing. In: 2013 IEEE international conference on image processing. IEEE, pp 2363\u20132367","DOI":"10.1109\/ICIP.2013.6738487"},{"issue":"12","key":"309_CR15","doi-asserted-by":"publisher","first-page":"1606","DOI":"10.1109\/TPAMI.2002.1114852","volume":"24","author":"JY Gil","year":"2002","unstructured":"Gil JY, Kimmel R (2002) Efficient dilation, erosion, opening, and closing algorithms. IEEE Trans Pattern Anal Mach Intell 24(12):1606\u20131617","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"309_CR16","unstructured":"Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. Adv Neural Inf Process Syst 27"},{"key":"309_CR17","doi-asserted-by":"publisher","first-page":"118631","DOI":"10.1016\/j.eswa.2022.118631","volume":"211","author":"C Guo","year":"2023","unstructured":"Guo C, Fan D, Jiang Z, Zhang D (2023) Mdfn: mask deep fusion network for visible and infrared image fusion without reference ground-truth. Expert Syst Appl 211:118631","journal-title":"Expert Syst Appl"},{"issue":"12","key":"309_CR18","first-page":"2341","volume":"33","author":"K He","year":"2010","unstructured":"He K, Sun J, Tang X (2010) Single image haze removal using dark channel prior. IEEE Trans Pattern Anal Mach Intell 33(12):2341\u20132353","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"309_CR19","doi-asserted-by":"publisher","unstructured":"Hide R (1977) Optics of the atmosphere: Scattering by molecules and particles. Phys Bulletin 521\u2013521. https:\/\/doi.org\/10.1088\/0031-9112\/28\/11\/025","DOI":"10.1088\/0031-9112\/28\/11\/025"},{"key":"309_CR20","doi-asserted-by":"crossref","unstructured":"Hu K, Zhang Q, Yuan M, Zhang Y (2024) Sfdfusion: An efficient spatial-frequency domain fusion network for infrared and visible image fusion. In: ECAI 2024. IOS Press, ???, pp 482\u2013489","DOI":"10.3233\/FAIA240524"},{"key":"309_CR21","doi-asserted-by":"publisher","first-page":"102147","DOI":"10.1016\/j.inffus.2023.102147","volume":"103","author":"H Li","year":"2024","unstructured":"Li H, Wu X-J (2024) Crossfuse: a novel cross attention mechanism based infrared and visible image fusion approach. Inf Fusion 103:102147","journal-title":"Inf Fusion"},{"key":"309_CR22","doi-asserted-by":"publisher","first-page":"107087","DOI":"10.1016\/j.knosys.2021.107087","volume":"224","author":"X Li","year":"2021","unstructured":"Li X, Zhou F, Tan H (2021) Joint image fusion and denoising via three-layer decomposition and sparse representation. Knowl-Based Syst 224:107087","journal-title":"Knowl-Based Syst"},{"key":"309_CR23","doi-asserted-by":"crossref","unstructured":"Li H, Fu Y (2024) Fcdfusion: a fast, low color deviation method for fusing visible and infrared image pairs. arXiv:2408.01080","DOI":"10.26599\/CVM.2025.9450330"},{"key":"309_CR24","doi-asserted-by":"publisher","unstructured":"Li B, Ren W, Fu D, Tao D, Feng D, Zeng W, Wang Z (2019) Benchmarking single-image dehazing and beyond. IEEE Trans on Image Process 492\u2013505. https:\/\/doi.org\/10.1109\/tip.2018.2867951","DOI":"10.1109\/tip.2018.2867951"},{"key":"309_CR25","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1016\/j.inffus.2014.09.004","volume":"24","author":"Y Liu","year":"2015","unstructured":"Liu Y, Liu S, Wang Z (2015) A general framework for image fusion based on multi-scale transform and sparse representation. Inf Fusion 24:147\u2013164","journal-title":"Inf Fusion"},{"issue":"12","key":"309_CR26","doi-asserted-by":"publisher","first-page":"1882","DOI":"10.1109\/LSP.2016.2618776","volume":"23","author":"Y Liu","year":"2016","unstructured":"Liu Y, Chen X, Ward RK, Wang ZJ (2016) Image fusion with convolutional sparse representation. IEEE Signal Process Lett 23(12):1882\u20131886","journal-title":"IEEE Signal Process Lett"},{"key":"309_CR27","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu Y, Chen X, Peng H, Wang Z (2017) Multi-focus image fusion with a deep convolutional neural network. Inf Fusion 36:191\u2013207","journal-title":"Inf Fusion"},{"key":"309_CR28","doi-asserted-by":"publisher","unstructured":"Liu Y, Chen X, Cheng J, Peng H, Wang Z (2018) Infrared and visible image fusion with convolutional neural networks. Int J Wavelets, Multiresol Inf Process 1850018. https:\/\/doi.org\/10.1142\/s0219691318500182","DOI":"10.1142\/s0219691318500182"},{"key":"309_CR29","doi-asserted-by":"crossref","unstructured":"Liu J, Fan X, Huang Z, Wu G, Liu R, Zhong W, Luo Z (2022) Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5802\u20135811","DOI":"10.1109\/CVPR52688.2022.00571"},{"key":"309_CR30","doi-asserted-by":"crossref","unstructured":"Li H, Wu X-J (2017) Multi-focus image fusion using dictionary learning and low-rank representation. In: Image and graphics: 9th International Conference, ICIG 2017, Shanghai, China, September 13-15, 2017, Revised Selected Papers, Part I 9. Springer, pp 675\u2013686","DOI":"10.1007\/978-3-319-71607-7_59"},{"key":"309_CR31","doi-asserted-by":"publisher","first-page":"128","DOI":"10.1016\/j.inffus.2020.11.009","volume":"69","author":"Y Long","year":"2021","unstructured":"Long Y, Jia H, Zhong Y, Jiang Y, Jia Y (2021) Rxdnfuse: a aggregated residual dense network for infrared and visible image fusion. Inf Fusion 69:128\u2013141","journal-title":"Inf Fusion"},{"key":"309_CR32","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","volume":"82","author":"J Ma","year":"2017","unstructured":"Ma J, Zhou Z, Wang B, Zong H (2017) Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys Technol 82:8\u201317","journal-title":"Infrared Phys Technol"},{"key":"309_CR33","first-page":"1","volume":"70","author":"J Ma","year":"2021","unstructured":"Ma J, Tang L, Xu M, Zhang H, Xiao G (2021) Stdfusionnet: an infrared and visible image fusion network based on salient target detection. IEEE Trans Instrum Meas 70:1\u201313","journal-title":"IEEE Trans Instrum Meas"},{"issue":"7","key":"309_CR34","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma J, Tang L, Fan F, Huang J, Mei X, Ma Y (2022) Swinfusion: cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA Journal of Automatica Sinica 9(7):1200\u20131217","journal-title":"IEEE\/CAA Journal of Automatica Sinica"},{"key":"309_CR35","doi-asserted-by":"publisher","unstructured":"Ma J, Yu W, Liang P, Li C, Jiang J (2019) Fusiongan: A generative adversarial network for infrared and visible image fusion. Inf Fusion 11\u201326. https:\/\/doi.org\/10.1016\/j.inffus.2018.09.004","DOI":"10.1016\/j.inffus.2018.09.004"},{"key":"309_CR36","doi-asserted-by":"publisher","unstructured":"Ma J, Zhou Z, Wang B, Zong H (2017) Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys Technol 8\u201317. https:\/\/doi.org\/10.1016\/j.infrared.2017.02.005","DOI":"10.1016\/j.infrared.2017.02.005"},{"issue":"7","key":"309_CR37","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1049\/el:20020212","volume":"38","author":"G Qu","year":"2002","unstructured":"Qu G, Zhang D, Yan P (2002) Information measure for performance of image fusion. Electron Lett 38(7):1","journal-title":"Electron Lett"},{"key":"309_CR38","unstructured":"Rajalingam B, Priya R (2018) Hybrid multimodality medical image fusion technique for feature enhancement in medical diagnosis. Int J Eng Sci Invention 2(Special issue):52\u201360"},{"issue":"4","key":"309_CR39","doi-asserted-by":"publisher","first-page":"355","DOI":"10.1088\/0957-0233\/8\/4\/002","volume":"8","author":"Y-J Rao","year":"1997","unstructured":"Rao Y-J (1997) In-fibre bragg grating sensors. Meas Sci Technol 8(4):355","journal-title":"Meas Sci Technol"},{"key":"309_CR40","doi-asserted-by":"crossref","unstructured":"Ren W, Liu S, Zhang H, Pan J, Cao X, Yang M-H (2016) Single image dehazing via multi-scale convolutional neural networks, pp 154\u2013169","DOI":"10.1007\/978-3-319-46475-6_10"},{"issue":"14","key":"309_CR41","doi-asserted-by":"publisher","first-page":"255","DOI":"10.1088\/0031-9155\/49\/14\/N07","volume":"49","author":"A Sassaroli","year":"2004","unstructured":"Sassaroli A, Fantini S (2004) Comment on the modified beer-lambert law for scattering media. Phys Med Biol 49(14):255","journal-title":"Phys Med Biol"},{"key":"309_CR42","doi-asserted-by":"publisher","first-page":"125472","DOI":"10.1016\/j.eswa.2024.125472","volume":"261","author":"W Song","year":"2025","unstructured":"Song W, Li Q, Gao M, Chehri A, Jeon G (2025) Sfinet: a semantic feature interactive learning network for full-time infrared and visible image fusion. Expert Syst Appl 261:125472","journal-title":"Expert Syst Appl"},{"key":"309_CR43","doi-asserted-by":"crossref","unstructured":"Tang L, Xiang X, Zhang H, Gong M, Ma J (2023) Divfusion: darkness-free infrared and visible image fusion. Inf Fusion 91:477\u2013493","DOI":"10.1016\/j.inffus.2022.10.034"},{"key":"309_CR44","doi-asserted-by":"crossref","unstructured":"Tang L, Yuan J, Zhang H, Jiang X, Ma J (2022) Piafusion: a progressive infrared and visible image fusion network based on illumination aware. Inf Fusion","DOI":"10.1016\/j.inffus.2022.03.007"},{"key":"309_CR45","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3216413","volume":"71","author":"Z Wang","year":"2022","unstructured":"Wang Z, Chen Y, Shao W, Li H, Zhang L (2022) Swinfuse: a residual swin transformer fusion network for infrared and visible images. IEEE Trans Instrum Meas 71:1\u201312","journal-title":"IEEE Trans Instrum Meas"},{"key":"309_CR46","doi-asserted-by":"publisher","first-page":"109192","DOI":"10.1016\/j.engappai.2024.109192","volume":"137","author":"C Wang","year":"2024","unstructured":"Wang C, Pu Y, Zhao Z, Nie R, Cao J, Xu D (2024) Fclfusion: a frequency-aware and collaborative learning for infrared and visible image fusion. Eng Appl Artif Intell 137:109192","journal-title":"Eng Appl Artif Intell"},{"key":"309_CR47","doi-asserted-by":"publisher","first-page":"107898","DOI":"10.1016\/j.engappai.2024.107898","volume":"132","author":"Y Wang","year":"2024","unstructured":"Wang Y, Pu J, Miao D, Zhang L, Zhang L, Du X (2024) Scgrfuse: an infrared and visible image fusion network based on spatial\/channel attention mechanism and gradient aggregation residual dense blocks. Eng Appl Artif Intell 132:107898","journal-title":"Eng Appl Artif Intell"},{"key":"309_CR48","doi-asserted-by":"crossref","unstructured":"Wang X, Chen X, Ren W, Han Z, Fan H, Tang Y, Liu L (2024) Compensation atmospheric scattering model and two-branch network for single image dehazing. IEEE Trans Emerg Topics Comput Intell","DOI":"10.1109\/TETCI.2024.3386838"},{"key":"309_CR49","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee J-Y, Kweon IS (2018) Cbam: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 3\u201319","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"309_CR50","doi-asserted-by":"crossref","unstructured":"Xu C, Li Q, Zhou Q, Jiang X, Yu D, Zhou Y (2022) Asymmetric cross-modal activation network for rgb-t salient object detection. Knowl-Based Syst 258:110047","DOI":"10.1016\/j.knosys.2022.110047"},{"key":"309_CR51","doi-asserted-by":"crossref","unstructured":"Xu H, Yuan J, Ma J (2023) Murf: mutually reinforcing multi-modal image registration and fusion. IEEE Trans Pattern Anal Mach Intell 45(10):12148\u201312166","DOI":"10.1109\/TPAMI.2023.3283682"},{"key":"309_CR52","doi-asserted-by":"publisher","first-page":"104693","DOI":"10.1016\/j.imavis.2023.104693","volume":"135","author":"G Yadav","year":"2023","unstructured":"Yadav G (2023) Yadav DK (2023) Contrast enhancement of region of interest of backlit image for surveillance systems based on multi-illumination fusion. Image Vision Comput 135:104693. https:\/\/doi.org\/10.1016\/j.imavis.2023.104693","journal-title":"Image Vision Comput"},{"key":"309_CR53","doi-asserted-by":"crossref","unstructured":"Yi X, Xu H, Zhang H, Tang L, Ma J (2024) Text-if: leveraging semantic text guidance for degradation-aware and interactive image fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","DOI":"10.1109\/CVPR52733.2024.02552"},{"key":"309_CR54","doi-asserted-by":"crossref","unstructured":"Yu M, Cui T, Lu H, Yue Y (2024) Vifnet: An end-to-end visible-infrared fusion network for image dehazing. Neurocomput 128105","DOI":"10.1016\/j.neucom.2024.128105"},{"key":"309_CR55","doi-asserted-by":"publisher","first-page":"227","DOI":"10.1016\/j.infrared.2017.05.007","volume":"83","author":"Y Zhang","year":"2017","unstructured":"Zhang Y, Zhang L, Bai X, Zhang L (2017) Infrared and visual image fusion through infrared feature extraction and visual information preservation. Infrared Phys Technol 83:227\u2013237","journal-title":"Infrared Phys Technol"},{"key":"309_CR56","doi-asserted-by":"publisher","first-page":"119442","DOI":"10.1016\/j.ins.2023.119442","volume":"647","author":"Z Zhang","year":"2023","unstructured":"Zhang Z, Liu Y, Xue W (2023) Ms-irtnet: multistage information interaction network for rgb-t semantic segmentation. Inf Sci 647:119442","journal-title":"Inf Sci"},{"key":"309_CR57","doi-asserted-by":"crossref","unstructured":"Zhang H, Tang L, Xiang X, Zuo X, Ma J (2024) Dispel darkness for better fusion: a controllable visual enhancer based on cross-modal conditional adversarial learning. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 26487\u201326496","DOI":"10.1109\/CVPR52733.2024.02501"},{"key":"309_CR58","doi-asserted-by":"crossref","unstructured":"Zhang H, Xu H, Xiao Y, Guo X, Ma J (2020) Rethinking the image fusion: a fast unified image fusion network based on proportional maintenance of gradient and intensity. In: Proceedings of the AAAI conference on artificial intelligence, vol. 34, pp 12797\u201312804","DOI":"10.1609\/aaai.v34i07.6975"},{"key":"309_CR59","doi-asserted-by":"publisher","unstructured":"Zhang Y, Zhang L, Bai X, Zhang L (2017) Infrared and visual image fusion through infrared feature extraction and visual information preservation. Infrared Phys Technol 227\u2013237. https:\/\/doi.org\/10.1016\/j.infrared.2017.05.007","DOI":"10.1016\/j.infrared.2017.05.007"},{"key":"309_CR60","doi-asserted-by":"crossref","unstructured":"Zhang H, Zuo X, Jiang J, Guo C, Ma J (2024) Mrfs: Mutually reinforcing image fusion and segmentation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 26974\u201326983","DOI":"10.1109\/CVPR52733.2024.02547"},{"key":"309_CR61","doi-asserted-by":"crossref","unstructured":"Zhao Z, Bai H, Zhu Y, Zhang J, Xu S, Zhang Y, Zhang K, Meng D, Timofte R, Van\u00a0Gool L (2023) Ddfm: denoising diffusion model for multi-modality image fusion. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 8082\u20138093","DOI":"10.1109\/ICCV51070.2023.00742"},{"key":"309_CR62","doi-asserted-by":"crossref","unstructured":"Zheng Y, Zhan J, He S, Dong J, Du Y (2023) Curricular contrastive regularization for physics-aware single image dehazing. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5785\u20135794","DOI":"10.1109\/CVPR52729.2023.00560"},{"key":"309_CR63","doi-asserted-by":"publisher","unstructured":"Zhou Z, Dong M, Xie X, Gao Z (2016) Fusion of infrared and visible images for night-vision context enhancement. Appl Optics 6480. https:\/\/doi.org\/10.1364\/ao.55.006480","DOI":"10.1364\/ao.55.006480"},{"key":"309_CR64","doi-asserted-by":"publisher","unstructured":"Zhou Z, Wang B, Li S, Dong M (2016) Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with gaussian and bilateral filters. Inf Fusion 15\u201326. https:\/\/doi.org\/10.1016\/j.inffus.2015.11.003","DOI":"10.1016\/j.inffus.2015.11.003"},{"issue":"11","key":"309_CR65","doi-asserted-by":"publisher","first-page":"3522","DOI":"10.1109\/TIP.2015.2446191","volume":"24","author":"Q Zhu","year":"2015","unstructured":"Zhu Q, Mai J, Shao L (2015) A fast single image haze removal algorithm using color attenuation prior. IEEE Trans Image Process 24(11):3522\u20133533","journal-title":"IEEE Trans Image Process"}],"container-title":["Journal of King Saud University Computer and Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-025-00309-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44443-025-00309-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-025-00309-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,19]],"date-time":"2025-11-19T15:46:07Z","timestamp":1763567167000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44443-025-00309-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,31]]},"references-count":65,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2025,11]]}},"alternative-id":["309"],"URL":"https:\/\/doi.org\/10.1007\/s44443-025-00309-7","relation":{},"ISSN":["1319-1578","2213-1248"],"issn-type":[{"type":"print","value":"1319-1578"},{"type":"electronic","value":"2213-1248"}],"subject":[],"published":{"date-parts":[[2025,10,31]]},"assertion":[{"value":"21 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 October 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interest"}}],"article-number":"290"}}