{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T09:26:11Z","timestamp":1776158771501,"version":"3.50.1"},"reference-count":57,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100018554","name":"Science and Technology Program of Gansu Province","doi-asserted-by":"publisher","award":["24YFFA024"],"award-info":[{"award-number":["24YFFA024"]}],"id":[{"id":"10.13039\/501100018554","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005229","name":"Key Science and Technology Foundation of Gansu Province","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100005229","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62061023"],"award-info":[{"award-number":["62061023"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004775","name":"Natural Science Foundation of Gansu Province","doi-asserted-by":"publisher","award":["24JRRA231"],"award-info":[{"award-number":["24JRRA231"]}],"id":[{"id":"10.13039\/501100004775","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Knowledge-Based Systems"],"published-print":{"date-parts":[[2026,5]]},"DOI":"10.1016\/j.knosys.2026.115861","type":"journal-article","created":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T23:50:02Z","timestamp":1774396202000},"page":"115861","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["SDD-Fuse: A multi-modality image fusion framework based on the spiking diffusion fusion model"],"prefix":"10.1016","volume":"341","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2206-828X","authenticated-orcid":false,"given":"Jing","family":"Di","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0002-8904-7384","authenticated-orcid":false,"given":"Heran","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3947-7215","authenticated-orcid":false,"given":"Jing","family":"Lian","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0005-1948-7909","authenticated-orcid":false,"given":"Shuhui","family":"Shi","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8303-2619","authenticated-orcid":false,"given":"Jizhao","family":"Liu","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.knosys.2026.115861_bib0001","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"5906","article-title":"CDDFUSE: correlation-driven dual-branch feature decomposition for multi-modality image fusion","author":"Zhao","year":"2023"},{"key":"10.1016\/j.knosys.2026.115861_bib0002","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2024.125427","article-title":"MixFuse: an iterative mix-attention transformer for multi-modal image fusion","volume":"261","author":"Li","year":"2025","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.knosys.2026.115861_bib0003","doi-asserted-by":"crossref","first-page":"115","DOI":"10.1002\/9781394234073.ch5","article-title":"Research in image processing for medical applications using the secure smart healthcare technique","author":"Modak","year":"2025","journal-title":"Des. Forecasting Models Dis. Manage."},{"key":"10.1016\/j.knosys.2026.115861_bib0004","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"26974","article-title":"Mrfs: mutually reinforcing image fusion and segmentation","author":"Zhang","year":"2024"},{"key":"10.1016\/j.knosys.2026.115861_bib0005","doi-asserted-by":"crossref","DOI":"10.1016\/j.compeleceng.2025.110133","article-title":"Cross-Modality target detection using infrared and visible image fusion for robust objection recognition","volume":"123","author":"Yu","year":"2025","journal-title":"Comput. Electr. Eng."},{"key":"10.1016\/j.knosys.2026.115861_bib0006","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.105819","article-title":"Anisotropic diffusion filter based fusion of NSST transformed medical images","volume":"90","author":"Gupta","year":"2024","journal-title":"Biomed. Signal Process. Contr."},{"key":"10.1016\/j.knosys.2026.115861_bib0007","doi-asserted-by":"crossref","DOI":"10.1016\/j.infrared.2023.104659","article-title":"Infrared and visible image fusion based on NSST and phase consistency adaptive DUAL channel PCNN","volume":"131","author":"Xie","year":"2023","journal-title":"Infrared Phys. Technol."},{"key":"10.1016\/j.knosys.2026.115861_bib0008","doi-asserted-by":"crossref","first-page":"5281","DOI":"10.1109\/TPAMI.2024.3368112","article-title":"A general spatial-frequency learning framework for multimodal image fusion","volume":"47","author":"Zhou","year":"2024","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.knosys.2026.115861_bib0009","doi-asserted-by":"crossref","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","article-title":"RFN-Nest: an end-to-end residual fusion network for infrared and visible images","volume":"73","author":"Li","year":"2021","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.knosys.2026.115861_bib0010","series-title":"European Conference on Computer Vision","first-page":"674","article-title":"FusionVAE: a deep hierarchical variational autoencoder for RGB image fusion","author":"Duffhauss","year":"2022"},{"key":"10.1016\/j.knosys.2026.115861_bib0011","first-page":"1","article-title":"Lenfusion: a joint low-light enhancement and fusion network for nighttime infrared and visible image fusion","volume":"73","author":"Chen","year":"2024","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.knosys.2026.115861_bib0012","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2024.111041","article-title":"MMAE: a universal image fusion method via mask attention mechanism","volume":"158","author":"Wang","year":"2025","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.knosys.2026.115861_bib0013","series-title":"Data Analytics for Intelligent Systems: Techniques and Solutions","first-page":"7","article-title":"An end-to-end content-aware generative adversarial network based method for multimodal medical image fusion","author":"Das","year":"2024"},{"key":"10.1016\/j.knosys.2026.115861_bib0014","series-title":"2024 4th International Conference on Computer Science and Blockchain (CCSB)","first-page":"466","article-title":"A lightweight GAN-based image fusion algorithm for visible and infrared images","author":"Wu","year":"2024"},{"key":"10.1016\/j.knosys.2026.115861_bib0015","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2024.107322","article-title":"Tq-cgan: a trible-generator quintuple-discriminator conditional generative adversarial network for multimodal grayscale medical image fusion","volume":"102","author":"Deng","year":"2025","journal-title":"Biomed. Signal Process. Contr."},{"key":"10.1016\/j.knosys.2026.115861_bib0016","doi-asserted-by":"crossref","DOI":"10.1016\/j.dsp.2023.104316","article-title":"DANT-GAN: a dual attention-based of nested training network for infrared and visible image fusion","volume":"145","author":"Li","year":"2024","journal-title":"Digit. Signal Process."},{"key":"10.1016\/j.knosys.2026.115861_bib0017","first-page":"8780","article-title":"Diffusion models beat gans on image synthesis","volume":"34","author":"Dhariwal","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115861_bib0018","first-page":"6840","article-title":"Denoising diffusion probabilistic models","volume":"33","author":"Ho","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115861_bib0019","series-title":"International Conference on Machine Learning","first-page":"8162","article-title":"Improved denoising diffusion probabilistic models","author":"Nichol","year":"2021"},{"key":"10.1016\/j.knosys.2026.115861_bib0020","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"10684","article-title":"High-resolution image synthesis with latent diffusion models","author":"Rombach","year":"2022"},{"issue":"1","key":"10.1016\/j.knosys.2026.115861_bib0021","doi-asserted-by":"crossref","first-page":"295","DOI":"10.1007\/s11760-023-02748-z","article-title":"An end-to-end based on semantic region guidance for infrared and visible image fusion","volume":"18","author":"Han","year":"2024","journal-title":"Signal Image Video Process."},{"key":"10.1016\/j.knosys.2026.115861_bib0022","article-title":"FusionDiff: multi-focus image fusion using denoising diffusion probabilistic models","volume":"238","author":"Li","year":"2024","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.knosys.2026.115861_bib0023","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"8082","article-title":"DDFM: Denoising diffusion model for multi-modality image fusion","author":"Zhao","year":"2023"},{"key":"10.1016\/j.knosys.2026.115861_bib0024","first-page":"39552","article-title":"Text-DiFuse: an interactive multi-modal image fusion framework based on text-modulated diffusion model","volume":"37","author":"Zhang","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115861_bib0025","doi-asserted-by":"crossref","first-page":"5705","DOI":"10.1109\/TIP.2023.3322046","article-title":"Dif-fusion: toward high color fidelity in infrared and visible image fusion with diffusion models","volume":"32","author":"Yue","year":"2023","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.knosys.2026.115861_bib0026","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102639","article-title":"LFDT-Fusion: A latent feature-guided diffusion transformer model for general image fusion","volume":"113","author":"Yang","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.knosys.2026.115861_bib0027","series-title":"Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision","first-page":"4912","article-title":"Spiking denoising diffusion probabilistic models","author":"Cao","year":"2024"},{"key":"10.1016\/j.knosys.2026.115861_bib0028","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","article-title":"FusionGAN: a generative adversarial network for infrared and visible image fusion","volume":"48","author":"Ma","year":"2019","journal-title":"Inf. fusion"},{"key":"10.1016\/j.knosys.2026.115861_bib0029","doi-asserted-by":"crossref","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","article-title":"DDcGAN: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion","volume":"29","author":"Ma","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.knosys.2026.115861_bib0030","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"5802","article-title":"Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection","author":"Liu","year":"2022"},{"key":"10.1016\/j.knosys.2026.115861_bib0031","first-page":"1","article-title":"GANMcC: a generative adversarial network with multiclassification constraints for infrared and visible image fusion","volume":"70","author":"Ma","year":"2020","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.knosys.2026.115861_bib0032","unstructured":"N. Skatchkovsky, O. Simeone, H. Jang, Learning to time-decode in spiking neural networks through the information bottleneck,(2021) arXiv: 2106.01177."},{"key":"10.1016\/j.knosys.2026.115861_bib0033","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"7059","article-title":"Fully spiking variational autoencoder","volume":"36","author":"Kamata","year":"2022"},{"key":"10.1016\/j.knosys.2026.115861_bib0034","doi-asserted-by":"crossref","DOI":"10.1016\/j.neunet.2024.106423","article-title":"Spiking generative adversarial network with attention scoring decoding","volume":"178","author":"Feng","year":"2024","journal-title":"Neural Netw."},{"key":"10.1016\/j.knosys.2026.115861_bib0035","series-title":"2024 5th International Conference on Computer, Big Data and Artificial Intelligence (ICCBD+ AI)","first-page":"627","article-title":"Spiking-diffusion: vector quantized discrete diffusion model with spiking neural networks","author":"Liu","year":"2024"},{"key":"10.1016\/j.knosys.2026.115861_bib0036","doi-asserted-by":"crossref","first-page":"132","DOI":"10.1109\/TAI.2024.3453229","article-title":"Spiking diffusion models","volume":"6","author":"Cao","year":"2024","journal-title":"IEEE Trans. Artif. Intell."},{"key":"10.1016\/j.knosys.2026.115861_bib0037","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"8239","article-title":"Diffusion models without attention","author":"Yan","year":"2024"},{"key":"10.1016\/j.knosys.2026.115861_bib0038","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"22428","article-title":"Smartbrush: text and shape guided object inpainting with diffusion model","author":"Xie","year":"2023"},{"key":"10.1016\/j.knosys.2026.115861_bib0039","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"10021","article-title":"Implicit diffusion models for continuous super-resolution","author":"Gao","year":"2023"},{"key":"10.1016\/j.knosys.2026.115861_bib0040","unstructured":"Z. Liu, Y. Wang, S. Vaidya, F. Ruehle, J. Halverson, M. Solja\u010di\u0107, T.Y. Hou, M. Tegmark, Kan: Kolmogorov-Arnold networks, (2024) arXiv: 2404.19756."},{"key":"10.1016\/j.knosys.2026.115861_bib0041","unstructured":"E. Hunsberger, C. Eliasmith, Spiking deep networks with LIF neurons,(2015) arXiv: 1510.08829."},{"key":"10.1016\/j.knosys.2026.115861_bib0042","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"5728","article-title":"Restormer: efficient transformer for high-resolution image restoration","author":"Zamir","year":"2022"},{"key":"10.1016\/j.knosys.2026.115861_bib0043","doi-asserted-by":"crossref","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","article-title":"PIAFusion: a progressive infrared and visible image fusion network based on illumination aware","volume":"83","author":"Tang","year":"2022","journal-title":"Inf. Fusion"},{"issue":"5","key":"10.1016\/j.knosys.2026.115861_bib0044","doi-asserted-by":"crossref","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","article-title":"DenseFuse: a fusion approach to infrared and visible images","volume":"28","author":"Li","year":"2018","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"10.1016\/j.knosys.2026.115861_bib0045","doi-asserted-by":"crossref","first-page":"108","DOI":"10.1097\/NPT.0b013e3182563795","article-title":"The whole brain atlas: www. med. harvard. edu\/aanlib","volume":"36","author":"Vidoni","year":"2012","journal-title":"J. Neurol. Phys. Ther."},{"key":"10.1016\/j.knosys.2026.115861_bib0046","doi-asserted-by":"crossref","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","article-title":"Image fusion meets deep learning: a survey and perspective","volume":"76","author":"Zhang","year":"2021","journal-title":"Inf. Fusion"},{"issue":"1","key":"10.1016\/j.knosys.2026.115861_bib0047","doi-asserted-by":"crossref","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","article-title":"U2Fusion: A unified unsupervised image fusion network","volume":"44","author":"Xu","year":"2020","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.knosys.2026.115861_bib0048","doi-asserted-by":"crossref","first-page":"80","DOI":"10.1016\/j.inffus.2022.11.010","article-title":"MUFusion: A general unsupervised image fusion network based on memory unit","volume":"92","author":"Cheng","year":"2023","journal-title":"Inf. Fusion"},{"issue":"7","key":"10.1016\/j.knosys.2026.115861_bib0049","doi-asserted-by":"crossref","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","article-title":"SwinFusion: cross-domain long-range learning for general image fusion via swin transformer","volume":"9","author":"Ma","year":"2022","journal-title":"IEEE\/CAA J. Autom. Sin."},{"key":"10.1016\/j.knosys.2026.115861_bib0050","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102414","article-title":"A general image fusion framework using multi-task semi-supervised learning","volume":"108","author":"Wang","year":"2024","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.knosys.2026.115861_bib0051","doi-asserted-by":"crossref","DOI":"10.1016\/j.optlaseng.2023.107925","article-title":"BTSFusion: Fusion of infrared and visible image via a mechanism of balancing texture and salience","volume":"173","author":"Qian","year":"2024","journal-title":"Opt. Lasers Eng."},{"issue":"4","key":"10.1016\/j.knosys.2026.115861_bib0052","doi-asserted-by":"crossref","DOI":"10.1016\/j.ipm.2024.103687","article-title":"FATFusion: A functional\u2013anatomical transformer for medical image fusion","volume":"61","author":"Tang","year":"2024","journal-title":"Inf. Process. Manage."},{"key":"10.1016\/j.knosys.2026.115861_bib0053","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2024.108381","article-title":"LrfNet: a real-time medical image fusion method guided by detail information","volume":"173","author":"He","year":"2024","journal-title":"Comput. Biol. Med."},{"key":"10.1016\/j.knosys.2026.115861_bib0054","doi-asserted-by":"crossref","first-page":"3550","DOI":"10.1109\/TIP.2024.3404660","article-title":"High-quality fusion and visualization for mr-pet brain tumor images via multi-dimensional features","volume":"33","author":"Wen","year":"2024","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.knosys.2026.115861_bib0055","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102450","article-title":"Diff-IF: multi-modality image fusion via diffusion model with fusion knowledge prior","volume":"110","author":"Yi","year":"2024","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.knosys.2026.115861_bib0056","unstructured":"G. Jocher, A. Chaurasia, A. Stoken, J. Borovec, Y. Kwon, K. Michael, J. Fang, C. Wong, Z. Yifu, D. Montes, et al., Ultralytics\/yolov5: v6. 2-yolov5 classification models, apple m1, reproducibility, clearml and deci. ai integrations, Zenodo (2022) 10.5281\/zenodo.7002879."},{"key":"10.1016\/j.knosys.2026.115861_bib0057","unstructured":"J. Song, C. Meng, S. Ermon, Denoising diffusion implicit models,(2020) arXiv: 2010.02502."}],"container-title":["Knowledge-Based Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0950705126005873?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0950705126005873?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T08:38:51Z","timestamp":1776155931000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0950705126005873"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,5]]},"references-count":57,"alternative-id":["S0950705126005873"],"URL":"https:\/\/doi.org\/10.1016\/j.knosys.2026.115861","relation":{},"ISSN":["0950-7051"],"issn-type":[{"value":"0950-7051","type":"print"}],"subject":[],"published":{"date-parts":[[2026,5]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"SDD-Fuse: A multi-modality image fusion framework based on the spiking diffusion fusion model","name":"articletitle","label":"Article Title"},{"value":"Knowledge-Based Systems","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.knosys.2026.115861","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"115861"}}