{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,19]],"date-time":"2026-01-19T00:50:13Z","timestamp":1768783813440,"version":"3.49.0"},"reference-count":57,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T00:00:00Z","timestamp":1751328000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T00:00:00Z","timestamp":1751328000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T00:00:00Z","timestamp":1751328000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T00:00:00Z","timestamp":1751328000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T00:00:00Z","timestamp":1751328000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T00:00:00Z","timestamp":1751328000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T00:00:00Z","timestamp":1751328000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100006407","name":"Henan Province Natural Science Foundation","doi-asserted-by":"publisher","award":["232300421093"],"award-info":[{"award-number":["232300421093"]}],"id":[{"id":"10.13039\/501100006407","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U22B2051"],"award-info":[{"award-number":["U22B2051"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62172371"],"award-info":[{"award-number":["62172371"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62102370"],"award-info":[{"award-number":["62102370"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62272422"],"award-info":[{"award-number":["62272422"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2021YFB3301500"],"award-info":[{"award-number":["2021YFB3301500"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Digital Signal Processing"],"published-print":{"date-parts":[[2025,7]]},"DOI":"10.1016\/j.dsp.2025.105136","type":"journal-article","created":{"date-parts":[[2025,3,12]],"date-time":"2025-03-12T12:19:54Z","timestamp":1741781994000},"page":"105136","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":2,"special_numbering":"C","title":["P3TFusion: Progressive Two-Stage Infrared and Visible Image Fusion Network Focused on Enhancing Target and Texture Information"],"prefix":"10.1016","volume":"162","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5823-2037","authenticated-orcid":false,"given":"Shupan","family":"Li","sequence":"first","affiliation":[]},{"given":"Yuan","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xiaoheng","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Mingyuan","family":"Jiu","sequence":"additional","affiliation":[]},{"given":"Yang","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Li","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Yan","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Ming","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Mingliang","family":"Xu","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.dsp.2025.105136_br0010","doi-asserted-by":"crossref","first-page":"185","DOI":"10.1016\/j.inffus.2022.09.019","article-title":"Current advances and future perspectives of image fusion: a comprehensive review","volume":"90","author":"Karim","year":"2023","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2025.105136_br0020","series-title":"Proceedings of the 30th ACM International Conference on Multimedia, MM '22","first-page":"4003","article-title":"DetFusion: a detection-driven infrared and visible image fusion network","author":"Sun","year":"2022"},{"key":"10.1016\/j.dsp.2025.105136_br0030","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.101806","article-title":"External multi-modal imaging sensor calibration for sensor fusion: a review","volume":"97","author":"Qiu","year":"2023","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2025.105136_br0040","doi-asserted-by":"crossref","first-page":"71","DOI":"10.1016\/j.optlaseng.2017.05.007","article-title":"A fusion algorithm for infrared and visible based on guided filtering and phase congruency in NSST domain","volume":"97","author":"Liu","year":"2017","journal-title":"Opt. Lasers Eng."},{"key":"10.1016\/j.dsp.2025.105136_br0050","first-page":"1","article-title":"A review of multimodal medical image fusion techniques","volume":"8","author":"Huang","year":"2020","journal-title":"Comput. Math. Methods Med."},{"key":"10.1016\/j.dsp.2025.105136_br0060","first-page":"21","article-title":"Medical image fusion method by deep learning","volume":"2","author":"Li","year":"2021","journal-title":"Int. J. Cogn. Comput. Eng."},{"key":"10.1016\/j.dsp.2025.105136_br0070","doi-asserted-by":"crossref","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","article-title":"Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network","volume":"82","author":"Tang","year":"2022","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2025.105136_br0080","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","article-title":"FusionGAN: a generative adversarial network for infrared and visible image fusion","volume":"48","author":"Ma","year":"2019","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2025.105136_br0090","doi-asserted-by":"crossref","first-page":"4733","DOI":"10.1109\/TIP.2020.2975984","article-title":"MDLatLRR: a novel decomposition method for infrared and visible image fusion","volume":"29","author":"Li","year":"2020","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"10.1016\/j.dsp.2025.105136_br0100","doi-asserted-by":"crossref","first-page":"1882","DOI":"10.1109\/LSP.2016.2618776","article-title":"Image fusion with convolutional sparse representation","volume":"23","author":"Liu","year":"2016","journal-title":"IEEE Signal Process. Lett."},{"issue":"5","key":"10.1016\/j.dsp.2025.105136_br0110","doi-asserted-by":"crossref","first-page":"743","DOI":"10.1109\/JSEN.2007.894926","article-title":"Region-based multimodal image fusion using ICA bases","volume":"7","author":"Cvejic","year":"2007","journal-title":"IEEE Sens. J."},{"key":"10.1016\/j.dsp.2025.105136_br0120","series-title":"Research on Infrared and Visible Image Fusion Based on Multi-level Information Extraction and Transmission","author":"li","year":"2023"},{"issue":"1","key":"10.1016\/j.dsp.2025.105136_br0130","article-title":"Deep learning-based image fusion: a survey","volume":"28","author":"Tang","year":"2023","journal-title":"J. Image Graph."},{"key":"10.1016\/j.dsp.2025.105136_br0140","series-title":"Optics, Photonics and Digital Technologies for Imaging Applications VII, vol. 12138","first-page":"35","article-title":"Multimodal super-resolution reconstruction based on encoder-decoder network","author":"Wang","year":"2022"},{"key":"10.1016\/j.dsp.2025.105136_br0150","doi-asserted-by":"crossref","first-page":"128","DOI":"10.1016\/j.inffus.2020.11.009","article-title":"RXDNFuse: a aggregated residual dense network for infrared and visible image fusion","volume":"69","author":"Long","year":"2020","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2025.105136_br0160","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36","first-page":"2126","article-title":"TransMEF: a transformer-based multi-exposure image fusion framework using self-supervised multi-task learning","author":"Qu","year":"2022"},{"issue":"12","key":"10.1016\/j.dsp.2025.105136_br0170","doi-asserted-by":"crossref","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","article-title":"NestFuse: an infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models","volume":"69","author":"Li","year":"2020","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"4","key":"10.1016\/j.dsp.2025.105136_br0180","first-page":"1514","article-title":"DIDFuse: deep image decomposition for infrared and visible image fusion","volume":"32","author":"Zhao","year":"2020","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"10.1016\/j.dsp.2025.105136_br0190","series-title":"National Conference on Artificial Intelligence","first-page":"12797","article-title":"Rethinking the image fusion: a fast unified image fusion network based on proportional maintenance of gradient and intensity","author":"Zhang","year":"2020"},{"issue":"1","key":"10.1016\/j.dsp.2025.105136_br0200","doi-asserted-by":"crossref","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","article-title":"U2Fusion: a unified unsupervised image fusion network","volume":"44","author":"Xu","year":"2022","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.dsp.2025.105136_br0210","series-title":"2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"19647","article-title":"RFNet: unsupervised network for mutually reinforcing multi-modal image registration and fusion","author":"Xu","year":"2022"},{"key":"10.1016\/j.dsp.2025.105136_br0220","doi-asserted-by":"crossref","first-page":"305","DOI":"10.1016\/j.inffus.2022.07.013","article-title":"UIFGAN: an unsupervised continual-learning generative adversarial network for unified image fusion","volume":"88","author":"Le","year":"2022","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2025.105136_br0230","doi-asserted-by":"crossref","first-page":"1383","DOI":"10.1109\/TMM.2020.2997127","article-title":"AttentionFGAN: infrared and visible image fusion using attention-based generative adversarial networks","volume":"23","author":"Li","year":"2021","journal-title":"IEEE Trans. Multimed."},{"issue":"5","key":"10.1016\/j.dsp.2025.105136_br0240","doi-asserted-by":"crossref","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","article-title":"DenseFuse: a fusion approach to infrared and visible images","volume":"28","author":"Li","year":"2019","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.dsp.2025.105136_br0250","doi-asserted-by":"crossref","first-page":"640","DOI":"10.1109\/TCI.2020.2965304","article-title":"VIF-Net: an unsupervised framework for infrared and visible image fusion","volume":"6","author":"Hou","year":"2020","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"10.1016\/j.dsp.2025.105136_br0260","doi-asserted-by":"crossref","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","article-title":"PIAFusion: a progressive infrared and visible image fusion network based on illumination aware","volume":"83\u201384","author":"Tang","year":"2022","journal-title":"Inf. Fusion"},{"issue":"4","key":"10.1016\/j.dsp.2025.105136_br0270","doi-asserted-by":"crossref","first-page":"532","DOI":"10.1109\/TCOM.1983.1095851","article-title":"The Laplacian Pyramid as a compact image code","volume":"31","author":"Burt","year":"1983","journal-title":"IEEE Trans. Commun."},{"issue":"7","key":"10.1016\/j.dsp.2025.105136_br0280","first-page":"674","article-title":"A theory for multiresolution signal decomposition: the wavelet representation","volume":"41","author":"Mallat","year":"1988","journal-title":"Commun. Pure Appl. Math."},{"issue":"10","key":"10.1016\/j.dsp.2025.105136_br0290","doi-asserted-by":"crossref","first-page":"3089","DOI":"10.1109\/TIP.2006.877507","article-title":"The nonsubsampled contourlet transform: theory, design, and applications","volume":"15","author":"Da","year":"2006","journal-title":"IEEE Trans. Image Process."},{"issue":"4","key":"10.1016\/j.dsp.2025.105136_br0300","first-page":"532","article-title":"An adaptive fusion algorithm for visible and infrared videos based on entropy and the cumulative distribution of gray levels","volume":"31","author":"Hu","year":"2017","journal-title":"IEEE Trans. Multimed."},{"issue":"1","key":"10.1016\/j.dsp.2025.105136_br0310","doi-asserted-by":"crossref","first-page":"10","DOI":"10.1016\/j.inffus.2010.04.001","article-title":"Pixel-level image fusion with simultaneous orthogonal matching pursuit","volume":"13","author":"Yang","year":"2012","journal-title":"Inf. Fusion"},{"issue":"21","key":"10.1016\/j.dsp.2025.105136_br0320","doi-asserted-by":"crossref","first-page":"21869","DOI":"10.1007\/s11042-017-4583-3","article-title":"An image fusion framework using novel dictionary based sparse representation","volume":"76","author":"Aishwarya","year":"2017","journal-title":"Multimed. Tools Appl."},{"issue":"5","key":"10.1016\/j.dsp.2025.105136_br0330","doi-asserted-by":"crossref","first-page":"1640","DOI":"10.1080\/01431161.2014.880819","article-title":"Infrared and visible image fusion based on random projection and sparse representation","volume":"35","author":"Wang","year":"2014","journal-title":"Int. J. Remote Sens."},{"key":"10.1016\/j.dsp.2025.105136_br0340","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1109\/TIM.2020.2986875","article-title":"Infrared and visible image fusion using visual saliency sparse representation and detail injection model","volume":"99","author":"Yang","year":"2020","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.dsp.2025.105136_br0350","doi-asserted-by":"crossref","first-page":"114","DOI":"10.1016\/j.infrared.2016.05.012","article-title":"Infrared and visible images fusion based on RPCA and NSCT","volume":"77","author":"Fu","year":"2016","journal-title":"Infrared Phys. Technol."},{"key":"10.1016\/j.dsp.2025.105136_br0360","series-title":"2013 6th International Congress on Image and Signal Processing (CISP)","first-page":"1046","article-title":"Image fusion based on non-negative matrix factorization and infrared feature extraction","author":"Mou","year":"2013"},{"key":"10.1016\/j.dsp.2025.105136_br0370","doi-asserted-by":"crossref","first-page":"100","DOI":"10.1016\/j.inffus.2016.02.001","article-title":"Infrared and visible image fusion via gradient transfer and total variation minimization","volume":"31","author":"Ma","year":"2016","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2025.105136_br0380","doi-asserted-by":"crossref","first-page":"52","DOI":"10.1016\/j.infrared.2016.01.009","article-title":"Two-scale image fusion of visible and infrared images using saliency detection","volume":"76","author":"Bavirisetti","year":"2016","journal-title":"Infrared Phys. Technol."},{"key":"10.1016\/j.dsp.2025.105136_br0390","first-page":"1","article-title":"STDFusionNet: an infrared and visible image fusion network based on salient target detection","volume":"70","author":"Ma","year":"2021","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.dsp.2025.105136_br0400","first-page":"4714","article-title":"DeepFuse: a deep unsupervised approach for exposure fusion with extreme exposure image pairs","author":"Prabhakar","year":"2017","journal-title":"IEEE Comput. Soc."},{"key":"10.1016\/j.dsp.2025.105136_br0410","doi-asserted-by":"crossref","first-page":"516","DOI":"10.1016\/j.ins.2017.09.010","article-title":"A novel multi-modality image fusion method based on image decomposition and sparse representation","volume":"432","author":"Zhu","year":"2018","journal-title":"Inf. Sci."},{"key":"10.1016\/j.dsp.2025.105136_br0420","doi-asserted-by":"crossref","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","article-title":"RFN-Nest: an end-to-end residual fusion network for infrared and visible images","volume":"73","author":"Li","year":"2021","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.dsp.2025.105136_br0430","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1109\/TIM.2020.3022438","article-title":"SEDRFuse: a symmetric encoder-decoder with residual block network for infrared and visible image fusion","volume":"70","author":"Jian","year":"2020","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.dsp.2025.105136_br0440","doi-asserted-by":"crossref","first-page":"824","DOI":"10.1109\/TCI.2021.3100986","article-title":"Classification saliency-based rule for visible and infrared image fusion","volume":"7","author":"Xu","year":"2021","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"10.1016\/j.dsp.2025.105136_br0450","series-title":"2020 25th International Conference on Pattern Recognition (ICPR)","first-page":"10675","article-title":"A dual-branch network for infrared and visible image fusion","author":"Fu","year":"2021"},{"key":"10.1016\/j.dsp.2025.105136_br0460","doi-asserted-by":"crossref","first-page":"103","DOI":"10.1016\/j.inffus.2022.10.010","article-title":"P2Sharpen: a progressive pansharpening network with deep spectral transformation","volume":"91","author":"Zhang","year":"2023","journal-title":"Inf. Fusion"},{"issue":"8","key":"10.1016\/j.dsp.2025.105136_br0470","doi-asserted-by":"crossref","first-page":"10535","DOI":"10.1109\/TPAMI.2023.3261282","article-title":"Visible and infrared image fusion using deep learning","volume":"45","author":"Zhang","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.dsp.2025.105136_br0480","doi-asserted-by":"crossref","first-page":"4070","DOI":"10.1109\/TIP.2021.3069339","article-title":"Different input resolutions and arbitrary output resolution: a meta learning-based deep framework for infrared and visible image fusion","volume":"30","author":"Li","year":"2021","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.dsp.2025.105136_br0490","series-title":"International Conference on Learning Representations","article-title":"An image is worth 16x16 words: transformers for image recognition at scale","author":"Dosovitskiy","year":"2021"},{"key":"10.1016\/j.dsp.2025.105136_br0500","series-title":"Proceedings of the European Conference on Computer Vision","first-page":"213","article-title":"End-to-end object detection with transformers","author":"Carion","year":"2020"},{"issue":"7","key":"10.1016\/j.dsp.2025.105136_br0510","doi-asserted-by":"crossref","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","article-title":"SwinFusion: cross-domain long-range learning for general image fusion via Swin transformer","volume":"9","author":"Ma","year":"2022","journal-title":"IEEE\/CAA J. Autom. Sin."},{"key":"10.1016\/j.dsp.2025.105136_br0520","doi-asserted-by":"crossref","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","article-title":"DDcGAN: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion","volume":"29","author":"Ma","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.dsp.2025.105136_br0530","article-title":"GANMcC: a generative adversarial network with multiclassification constraints for infrared and visible image fusion","volume":"70","author":"Ma","year":"2020","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.dsp.2025.105136_br0540","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2022.116905","article-title":"DSG-fusion: infrared and visible image fusion via generative adversarial networks and guided filter","volume":"200","author":"Yang","year":"2022","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.dsp.2025.105136_br0550","series-title":"Feedback Circuits and Op. Amps, Tutorial Guides in Electronic Engineering","author":"Horrocks","year":"2013"},{"key":"10.1016\/j.dsp.2025.105136_br0560","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"3496","article-title":"LLVIP: a visible-infrared paired dataset for low-light vision","author":"Jia","year":"2021"},{"key":"10.1016\/j.dsp.2025.105136_br0570","doi-asserted-by":"crossref","first-page":"249","DOI":"10.1016\/j.dib.2017.09.038","article-title":"The TNO multiband image data collection","volume":"15","author":"Toet","year":"2017","journal-title":"Data Brief"}],"container-title":["Digital Signal Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1051200425001587?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1051200425001587?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T15:06:45Z","timestamp":1762355205000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1051200425001587"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7]]},"references-count":57,"alternative-id":["S1051200425001587"],"URL":"https:\/\/doi.org\/10.1016\/j.dsp.2025.105136","relation":{},"ISSN":["1051-2004"],"issn-type":[{"value":"1051-2004","type":"print"}],"subject":[],"published":{"date-parts":[[2025,7]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"P3TFusion: Progressive Two-Stage Infrared and Visible Image Fusion Network Focused on Enhancing Target and Texture Information","name":"articletitle","label":"Article Title"},{"value":"Digital Signal Processing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.dsp.2025.105136","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Inc. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"105136"}}