{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T15:56:24Z","timestamp":1775318184145,"version":"3.50.1"},"reference-count":58,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100007839","name":"Yunnan University","doi-asserted-by":"publisher","award":["KC-24249476"],"award-info":[{"award-number":["KC-24249476"]}],"id":[{"id":"10.13039\/501100007839","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62162065"],"award-info":[{"award-number":["62162065"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Digital Signal Processing"],"published-print":{"date-parts":[[2026,6]]},"DOI":"10.1016\/j.dsp.2026.105991","type":"journal-article","created":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T00:31:40Z","timestamp":1770769900000},"page":"105991","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["MWEFusion: Multi-scale wavelet-based feature enhancement network for infrared and visible image fusion"],"prefix":"10.1016","volume":"176","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-2039-1382","authenticated-orcid":false,"given":"Yulong","family":"Li","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0009-7003-9377","authenticated-orcid":false,"given":"Zheng","family":"Guan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6674-8140","authenticated-orcid":false,"given":"Xue","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-5688-704X","authenticated-orcid":false,"given":"Hao","family":"Dai","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.dsp.2026.105991_bib0001","doi-asserted-by":"crossref","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","article-title":"Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network","volume":"82","author":"Tang","year":"2022","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0002","doi-asserted-by":"crossref","first-page":"182","DOI":"10.1016\/j.compind.2018.12.013","article-title":"A novel convolutional neural network based fault recognition method via image fusion of multi-vibration-signals","volume":"105","author":"Wang","year":"2019","journal-title":"Comput. Ind."},{"key":"10.1016\/j.dsp.2026.105991_bib0003","first-page":"1","article-title":"Lenfusion: a joint low-light enhancement and fusion network for nighttime infrared and visible image fusion","volume":"73","author":"Chen","year":"2024","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"12","key":"10.1016\/j.dsp.2026.105991_bib0004","doi-asserted-by":"crossref","first-page":"2789","DOI":"10.3390\/rs14122789","article-title":"Infrared and visible image fusion with deep neural network in enhanced flight vision system","volume":"14","author":"Gao","year":"2022","journal-title":"Remote Sens."},{"key":"10.1016\/j.dsp.2026.105991_bib0005","doi-asserted-by":"crossref","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","article-title":"Infrared and visible image fusion methods and applications: a survey","volume":"45","author":"Ma","year":"2019","journal-title":"Inf. fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0006","doi-asserted-by":"crossref","DOI":"10.1016\/j.compag.2021.106491","article-title":"Recent advances in image fusion technology in agriculture","volume":"191","author":"Li","year":"2021","journal-title":"Comput. Electron. Agric."},{"key":"10.1016\/j.dsp.2026.105991_bib0007","doi-asserted-by":"crossref","first-page":"1818","DOI":"10.1109\/LSP.2021.3109818","article-title":"Smoa: searching a modality-oriented architecture for infrared and visible image fusion","volume":"28","author":"Liu","year":"2021","journal-title":"IEEE Signal Process. Lett."},{"key":"10.1016\/j.dsp.2026.105991_bib0008","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.101828","article-title":"An interactively reinforced paradigm for joint infrared-visible image fusion and saliency object detection","volume":"98","author":"Wang","year":"2023","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0009","doi-asserted-by":"crossref","first-page":"174","DOI":"10.1016\/j.inffus.2022.12.022","article-title":"A perceptual framework for infrared\u2013visible image fusion based on multiscale structure decomposition and biological vision","volume":"93","author":"Zhou","year":"2023","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0010","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"2760","article-title":"Image fusion with local spectral consistency and dynamic gradient sparsity","author":"Chen","year":"2014"},{"key":"10.1016\/j.dsp.2026.105991_bib0011","doi-asserted-by":"crossref","first-page":"4733","DOI":"10.1109\/TIP.2020.2975984","article-title":"MDLatLRR: a novel decomposition method for infrared and visible image fusion","volume":"29","author":"Li","year":"2020","journal-title":"IEEE Trans. Image Process."},{"issue":"4","key":"10.1016\/j.dsp.2026.105991_bib0012","doi-asserted-by":"crossref","first-page":"624","DOI":"10.1109\/TMM.2009.2017640","article-title":"Segmentation-driven image fusion based on alpha-stable modeling of wavelet coefficients","volume":"11","author":"Wan","year":"2009","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.dsp.2026.105991_bib0013","doi-asserted-by":"crossref","DOI":"10.1016\/j.cviu.2022.103407","article-title":"CUFD: an encoder\u2013decoder network for visible and infrared image fusion based on common and unique feature decomposition","volume":"218","author":"Xu","year":"2022","journal-title":"Comput. Vision Image Understand."},{"key":"10.1016\/j.dsp.2026.105991_bib0014","series-title":"European Conference on Computer Vision","first-page":"719","article-title":"Fusion from decomposition: a self-supervised decomposition approach for image fusion","author":"Liang","year":"2022"},{"key":"10.1016\/j.dsp.2026.105991_bib0015","series-title":"European Conference on Computer Vision","first-page":"740","article-title":"Microsoft coco: common objects in context","author":"Lin","year":"2014"},{"key":"10.1016\/j.dsp.2026.105991_bib0016","first-page":"1","article-title":"GANMcC: a generative adversarial network with multiclassification constraints for infrared and visible image fusion","volume":"70","author":"Ma","year":"2020","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"10","key":"10.1016\/j.dsp.2026.105991_bib0017","doi-asserted-by":"crossref","first-page":"2761","DOI":"10.1007\/s11263-021-01501-8","article-title":"SDNet: a versatile squeeze-and-decomposition network for real-time image fusion","volume":"129","author":"Zhang","year":"2021","journal-title":"Int. J. Comput. Vis."},{"issue":"1","key":"10.1016\/j.dsp.2026.105991_bib0018","doi-asserted-by":"crossref","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","article-title":"U2Fusion: a unified unsupervised image fusion network","volume":"44","author":"Xu","year":"2020","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.dsp.2026.105991_bib0019","doi-asserted-by":"crossref","first-page":"769","DOI":"10.1109\/TCI.2023.3304471","article-title":"Contrast saliency information guided infrared and visible image fusion","volume":"9","author":"Wang","year":"2023","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"10.1016\/j.dsp.2026.105991_bib0020","article-title":"STFuse: infrared and visible image fusion via semisupervised transfer learning","author":"Wang","year":"2023","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"10.1016\/j.dsp.2026.105991_bib0021","doi-asserted-by":"crossref","first-page":"5977","DOI":"10.1109\/TMM.2025.3565970","article-title":"PID controller-driven network for image fusion","volume":"27","author":"Wang","year":"2025","journal-title":"IEEE Trans. Multimed."},{"issue":"3","key":"10.1016\/j.dsp.2026.105991_bib0022","doi-asserted-by":"crossref","first-page":"1186","DOI":"10.1109\/TCSVT.2021.3075745","article-title":"Efficient and model-based infrared and visible image fusion via algorithm unrolling","volume":"32","author":"Zhao","year":"2021","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.dsp.2026.105991_bib0023","doi-asserted-by":"crossref","DOI":"10.1016\/j.optcom.2024.131024","article-title":"Wavefusionnet: infrared and visible image fusion based on multi-scale feature encoder\u2013decoder and discrete wavelet decomposition","volume":"573","author":"Liu","year":"2024","journal-title":"Opt. Commun."},{"key":"10.1016\/j.dsp.2026.105991_bib0024","first-page":"1","article-title":"DePF: a novel fusion approach based on decomposition pooling for infrared and visible images","volume":"72","author":"Li","year":"2023","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"8","key":"10.1016\/j.dsp.2026.105991_bib0025","doi-asserted-by":"crossref","first-page":"10535","DOI":"10.1109\/TPAMI.2023.3261282","article-title":"Visible and infrared image fusion using deep learning","volume":"45","author":"Zhang","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.dsp.2026.105991_bib0026","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.102039","article-title":"Cs2fusion: contrastive learning for self-supervised infrared and visible image fusion by estimating feature compensation map","volume":"102","author":"Wang","year":"2024","journal-title":"Inf. Fusion"},{"issue":"5","key":"10.1016\/j.dsp.2026.105991_bib0027","doi-asserted-by":"crossref","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","article-title":"DenseFuse: a fusion approach to infrared and visible images","volume":"28","author":"Li","year":"2018","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"10.1016\/j.dsp.2026.105991_bib0028","doi-asserted-by":"crossref","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","article-title":"NestFuse: an infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models","volume":"69","author":"Li","year":"2020","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.dsp.2026.105991_bib0029","doi-asserted-by":"crossref","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","article-title":"RFN-Nest: an end-to-end residual fusion network for infrared and visible images","volume":"73","author":"Li","year":"2021","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0030","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"5906","article-title":"Cddfuse: correlation-driven dual-branch feature decomposition for multi-modality image fusion","author":"Zhao","year":"2023"},{"key":"10.1016\/j.dsp.2026.105991_bib0031","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","article-title":"FusionGAN: a generative adversarial network for infrared and visible image fusion","volume":"48","author":"Ma","year":"2019","journal-title":"Inf. fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0032","doi-asserted-by":"crossref","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","article-title":"DDcGAN: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion","volume":"29","author":"Ma","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.dsp.2026.105991_bib0033","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"5802","article-title":"Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection","author":"Liu","year":"2022"},{"key":"10.1016\/j.dsp.2026.105991_bib0034","doi-asserted-by":"crossref","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","article-title":"IFCNN: a general image fusion framework based on convolutional neural network","volume":"54","author":"Zhang","year":"2020","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0035","first-page":"1","article-title":"Correlation-guided discriminative cross-modality features network for infrared and visible image fusion","volume":"73","author":"Cai","year":"2023","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.dsp.2026.105991_bib0036","doi-asserted-by":"crossref","first-page":"205","DOI":"10.1016\/j.inffus.2022.09.030","article-title":"SGFusion: a saliency guided deep-learning framework for pixel-level image fusion","volume":"91","author":"Liu","year":"2023","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0037","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102790","article-title":"Textfusion: unveiling the power of textual semantics for controllable image fusion","volume":"117","author":"Cheng","year":"2025","journal-title":"Information Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0038","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops","first-page":"773","article-title":"Multi-level wavelet-CNN for image restoration","author":"Liu","year":"2018"},{"issue":"12","key":"10.1016\/j.dsp.2026.105991_bib0039","doi-asserted-by":"crossref","first-page":"15979","DOI":"10.1109\/TPAMI.2023.3307666","article-title":"Wavelet approximation-aware residual network for single image deraining","volume":"45","author":"Hsu","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.dsp.2026.105991_bib0040","series-title":"Wavelets and Statistics","first-page":"281","article-title":"The stationary wavelet transform and some statistical applications","author":"Nason","year":"1995"},{"issue":"10","key":"10.1016\/j.dsp.2026.105991_bib0041","doi-asserted-by":"crossref","first-page":"2464","DOI":"10.1109\/78.157290","article-title":"The discrete wavelet transform: wedding the a trous and mallat algorithms","volume":"40","author":"Shensa","year":"2002","journal-title":"IEEE Trans. Signal Process."},{"key":"10.1016\/j.dsp.2026.105991_bib0042","article-title":"Infrared-visible image fusion using dual-branch auto-encoder with invertible high-frequency encoding","author":"Liu","year":"2024","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.dsp.2026.105991_bib0043","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2025.111594","article-title":"Haarfuse: a dual-branch infrared and visible light image fusion network based on haar wavelet transform","volume":"164","author":"Wang","year":"2025","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.dsp.2026.105991_bib0044","doi-asserted-by":"crossref","first-page":"4776","DOI":"10.1109\/TMM.2023.3326296","article-title":"CAMF: an interpretable infrared and visible image fusion network based on class activation mapping","volume":"26","author":"Tang","year":"2023","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.dsp.2026.105991_bib0045","series-title":"Proceedings of the 31st ACM International Conference on Multimedia","first-page":"4471","article-title":"Learning a graph neural network with cross modality interaction for image fusion","author":"Li","year":"2023"},{"issue":"9","key":"10.1016\/j.dsp.2026.105991_bib0046","doi-asserted-by":"crossref","first-page":"11040","DOI":"10.1109\/TPAMI.2023.3268209","article-title":"Lrrnet: a novel representation learning guided fusion network for infrared and visible images","volume":"45","author":"Li","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.dsp.2026.105991_bib0047","doi-asserted-by":"crossref","first-page":"477","DOI":"10.1016\/j.inffus.2022.10.034","article-title":"DIVFusion: darkness-free infrared and visible image fusion","volume":"91","author":"Tang","year":"2023","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0048","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2025.102931","article-title":"A degradation-aware guided fusion network for infrared and visible image","volume":"118","author":"Wang","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0049","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102655","article-title":"DDBFusion: an unified image decomposition and fusion framework based on dual decomposition and B\u00e9zier curves","volume":"114","author":"Zhang","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0050","series-title":"Digital Image Processing","author":"Gonzalez","year":"2009"},{"issue":"4","key":"10.1016\/j.dsp.2026.105991_bib0051","doi-asserted-by":"crossref","first-page":"308","DOI":"10.1049\/el:20000267","article-title":"Objective image fusion performance measure","volume":"36","author":"Xydeas","year":"2000","journal-title":"Electron. Lett."},{"key":"10.1016\/j.dsp.2026.105991_bib0052","series-title":"APS March Meeting Abstracts","first-page":"H1","article-title":"Image quality metric based on mutual information of image features","volume":"2014","author":"Haghighat","year":"2014"},{"issue":"3","key":"10.1016\/j.dsp.2026.105991_bib0053","doi-asserted-by":"crossref","first-page":"224","DOI":"10.1016\/j.inffus.2006.10.004","article-title":"Guest editorial: concurrent learning and fusion","volume":"8","author":"Perlovsky","year":"2007","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2026.105991_bib0054","article-title":"Enhanced deep learning and quantum variational classifier for large-scale data analysis","author":"Sudha","year":"2025","journal-title":"Knowl. Based Syst."},{"key":"10.1016\/j.dsp.2026.105991_bib0055","doi-asserted-by":"crossref","DOI":"10.1016\/j.aei.2025.103622","article-title":"Quantum generative adversarial network based on the quantum born machine","volume":"68","author":"Ding","year":"2025","journal-title":"Adv. Eng. Inf."},{"issue":"2","key":"10.1016\/j.dsp.2026.105991_bib0056","doi-asserted-by":"crossref","first-page":"23","DOI":"10.1007\/s12293-025-00463-5","article-title":"Image denoising with hybrid classical-Quantum convolutional neural network: X. He et al","volume":"17","author":"He","year":"2025","journal-title":"Memetic Comput."},{"key":"10.1016\/j.dsp.2026.105991_bib0057","doi-asserted-by":"crossref","first-page":"4784","DOI":"10.1109\/JSTARS.2024.3362748","article-title":"Fsod4rsi: few-shot object detection for remote sensing images via features aggregation and scale attention","volume":"17","author":"Gao","year":"2024","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"10.1016\/j.dsp.2026.105991_bib0058","article-title":"Swinbtc: transfer learning to brain tumor classification for healthcare electronics using augmented MR images","author":"Gao","year":"2025","journal-title":"IEEE Trans. Consum. Electron."}],"container-title":["Digital Signal Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1051200426001107?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1051200426001107?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T15:00:16Z","timestamp":1775314816000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1051200426001107"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,6]]},"references-count":58,"alternative-id":["S1051200426001107"],"URL":"https:\/\/doi.org\/10.1016\/j.dsp.2026.105991","relation":{},"ISSN":["1051-2004"],"issn-type":[{"value":"1051-2004","type":"print"}],"subject":[],"published":{"date-parts":[[2026,6]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"MWEFusion: Multi-scale wavelet-based feature enhancement network for infrared and visible image fusion","name":"articletitle","label":"Article Title"},{"value":"Digital Signal Processing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.dsp.2026.105991","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Published by Elsevier Inc.","name":"copyright","label":"Copyright"}],"article-number":"105991"}}