{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T05:43:33Z","timestamp":1776145413659,"version":"3.50.1"},"reference-count":69,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100019033","name":"Key Research and Development Program of Liaoning Province","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100019033","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100021171","name":"Basic and Applied Basic Research Foundation of Guangdong Province","doi-asserted-by":"publisher","award":["2023A1515140132"],"award-info":[{"award-number":["2023A1515140132"]}],"id":[{"id":"10.13039\/501100021171","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100013142","name":"Key Research and Development Project of Hainan Province","doi-asserted-by":"publisher","award":["2023JH2\/101800013"],"award-info":[{"award-number":["2023JH2\/101800013"]}],"id":[{"id":"10.13039\/501100013142","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Biomedical Signal Processing and Control"],"published-print":{"date-parts":[[2026,7]]},"DOI":"10.1016\/j.bspc.2026.110075","type":"journal-article","created":{"date-parts":[[2026,3,17]],"date-time":"2026-03-17T10:56:16Z","timestamp":1773744976000},"page":"110075","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"PB","title":["A multi-scale expert gating network with cross-modal attention for multimodal medical image fusion"],"prefix":"10.1016","volume":"120","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6501-4097","authenticated-orcid":false,"given":"Hegui","family":"Zhu","sequence":"first","affiliation":[]},{"given":"Qi","family":"Cui","sequence":"additional","affiliation":[]},{"given":"Jiayi","family":"Li","sequence":"additional","affiliation":[]},{"given":"Libo","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"issue":"17","key":"10.1016\/j.bspc.2026.110075_b1","doi-asserted-by":"crossref","first-page":"26731","DOI":"10.1007\/s11042-022-14305-w","article-title":"Machine learning and deep learning approach for medical image analysis: diagnosis to detection","volume":"82","author":"Rana","year":"2023","journal-title":"Multimedia Tools Appl."},{"issue":"2","key":"10.1016\/j.bspc.2026.110075_b2","doi-asserted-by":"crossref","first-page":"114","DOI":"10.1016\/j.inffus.2009.05.003","article-title":"MRI and PET image fusion by combining IHS and retina-inspired models","volume":"11","author":"Daneshvar","year":"2010","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.bspc.2026.110075_b3","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2021.105005","article-title":"A multiscale double-branch residual attention network for anatomical\u2013functional medical image fusion","volume":"141","author":"Li","year":"2022","journal-title":"Comput. Biol. Med."},{"issue":"12","key":"10.1016\/j.bspc.2026.110075_b4","doi-asserted-by":"crossref","first-page":"5855","DOI":"10.1109\/TIP.2017.2745202","article-title":"Anatomical-functional image fusion by information of interest in local Laplacian filtering domain","volume":"26","author":"Du","year":"2017","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.bspc.2026.110075_b5","series-title":"2016 International Conference on Signal Processing, Communication, Power and Embedded System","first-page":"1089","article-title":"PCA-DWT based medical image fusion using non sub-sampled contourlet transform","author":"Madanala","year":"2016"},{"key":"10.1016\/j.bspc.2026.110075_b6","doi-asserted-by":"crossref","first-page":"625","DOI":"10.1016\/j.procs.2015.10.057","article-title":"Multi-modality medical image fusion using discrete wavelet transform","volume":"70","author":"Bhavana","year":"2015","journal-title":"Procedia Comput. Sci."},{"issue":"3","key":"10.1016\/j.bspc.2026.110075_b7","article-title":"A novel method of multimodal medical image fusion based on hybrid approach of NSCT and DTCWT","volume":"18","author":"Alseelawi","year":"2022","journal-title":"Int. J. Online & Biomed. Eng."},{"issue":"5","key":"10.1016\/j.bspc.2026.110075_b8","first-page":"59","article-title":"Multimodal medical image fusion based on the VGG19 model in the NSCT domain","volume":"17","author":"Liu","year":"2024","journal-title":"Recent. Adv. Comput. Sci. Commun. (Formerly: Recent. Patents Comput. Science)"},{"issue":"3","key":"10.1016\/j.bspc.2026.110075_b9","doi-asserted-by":"crossref","first-page":"2295","DOI":"10.1007\/s11277-023-10542-w","article-title":"A novel MRI and PET image fusion in the NSST domain using YUV color space based on convolutional neural networks","volume":"131","author":"Sebastian","year":"2023","journal-title":"Wirel. Pers. Commun."},{"key":"10.1016\/j.bspc.2026.110075_b10","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.105819","article-title":"Anisotropic diffusion filter based fusion of NSST transformed medical images","volume":"90","author":"Gupta","year":"2024","journal-title":"Biomed. Signal Process. Control."},{"key":"10.1016\/j.bspc.2026.110075_b11","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2021.102788","article-title":"Multi-modal medical image fusion framework using co-occurrence filter and local extrema in NSST domain","volume":"68","author":"Diwakar","year":"2021","journal-title":"Biomed. Signal Process. Control."},{"issue":"1","key":"10.1016\/j.bspc.2026.110075_b12","article-title":"Clustering based multi-modality medical image fusion","volume":"1478","author":"Dhaundiyal","year":"2020","journal-title":"J. Phys.: Conf. Ser."},{"issue":"9","key":"10.1016\/j.bspc.2026.110075_b13","doi-asserted-by":"crossref","first-page":"2629","DOI":"10.1007\/s11517-024-03089-w","article-title":"Multi-modal medical image fusion using improved dual-channel PCNN","volume":"62","author":"Sinha","year":"2024","journal-title":"Med. Biol. Eng. Comput."},{"key":"10.1016\/j.bspc.2026.110075_b14","first-page":"1","article-title":"Multimodal medical image fusion using the MBM-PCNN model","author":"Koteswara Rao","year":"2025","journal-title":"IETE J. Res."},{"key":"10.1016\/j.bspc.2026.110075_b15","doi-asserted-by":"crossref","first-page":"177","DOI":"10.1016\/j.inffus.2021.06.001","article-title":"EMFusion: An unsupervised enhanced medical image fusion network","volume":"76","author":"Xu","year":"2021","journal-title":"Inf. Fusion"},{"issue":"1","key":"10.1016\/j.bspc.2026.110075_b16","doi-asserted-by":"crossref","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","article-title":"U2Fusion: A unified unsupervised image fusion network","volume":"44","author":"Xu","year":"2022","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.bspc.2026.110075_b17","doi-asserted-by":"crossref","first-page":"80","DOI":"10.1016\/j.inffus.2022.11.010","article-title":"MUFusion: A general unsupervised image fusion network based on memory unit","volume":"92","author":"Cheng","year":"2023","journal-title":"Inf. Fusion"},{"issue":"7","key":"10.1016\/j.bspc.2026.110075_b18","doi-asserted-by":"crossref","first-page":"5921","DOI":"10.1109\/TCSVT.2023.3342808","article-title":"Learning to search a lightweight generalized network for medical image fusion","volume":"34","author":"Mu","year":"2024","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"4","key":"10.1016\/j.bspc.2026.110075_b19","doi-asserted-by":"crossref","DOI":"10.1016\/j.ipm.2024.103687","article-title":"FATFusion: A functional\u2013anatomical transformer for medical image fusion","volume":"61","author":"Tang","year":"2024","journal-title":"Inf. Process. Manage."},{"issue":"5","key":"10.1016\/j.bspc.2026.110075_b20","doi-asserted-by":"crossref","first-page":"3317","DOI":"10.1109\/JBHI.2024.3391620","article-title":"MACTFusion: Lightweight cross transformer for adaptive multimodal medical image fusion","volume":"29","author":"Xie","year":"2025","journal-title":"IEEE J. Biomed. Health Informatics"},{"issue":"1","key":"10.1016\/j.bspc.2026.110075_b21","doi-asserted-by":"crossref","first-page":"339","DOI":"10.1007\/s10278-022-00696-7","article-title":"U-Patch GAN: A medical image fusion method based on GAN","volume":"36","author":"Fan","year":"2023","journal-title":"J. Digit. Imaging"},{"key":"10.1016\/j.bspc.2026.110075_b22","doi-asserted-by":"crossref","first-page":"134","DOI":"10.1016\/j.inffus.2022.10.017","article-title":"GAN review: Models and medical image fusion applications","volume":"91","author":"Zhou","year":"2023","journal-title":"Inf. Fusion"},{"issue":"4","key":"10.1016\/j.bspc.2026.110075_b23","doi-asserted-by":"crossref","first-page":"259","DOI":"10.1016\/S1566-2535(03)00046-0","article-title":"A general framework for multiresolution image fusion: from pixels to regions","volume":"4","author":"Piella","year":"2003","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.bspc.2026.110075_b24","series-title":"2017 20th International Conference on Information Fusion","first-page":"1","article-title":"A medical image fusion method based on convolutional neural networks","author":"Liu","year":"2017"},{"key":"10.1016\/j.bspc.2026.110075_b25","series-title":"An image is worth 16x16 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020"},{"key":"10.1016\/j.bspc.2026.110075_b26","series-title":"Multi-scale context aggregation by dilated convolutions","author":"Yu","year":"2015"},{"key":"10.1016\/j.bspc.2026.110075_b27","series-title":"2017 IEEE Conference on Computer Vision and Pattern Recognition","first-page":"5987","article-title":"Aggregated residual transformations for deep neural networks","author":"Xie","year":"2017"},{"key":"10.1016\/j.bspc.2026.110075_b28","series-title":"Outrageously large neural networks: The sparsely-gated mixture-of-experts layer","author":"Shazeer","year":"2017"},{"issue":"3","key":"10.1016\/j.bspc.2026.110075_b29","doi-asserted-by":"crossref","first-page":"1712","DOI":"10.1109\/TCSVT.2023.3296745","article-title":"TUFusion: A transformer-based universal fusion algorithm for multimodal images","volume":"34","author":"Zhao","year":"2024","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.bspc.2026.110075_b30","doi-asserted-by":"crossref","first-page":"195","DOI":"10.1016\/j.bspc.2017.02.005","article-title":"Medical image fusion based on sparse representation of classified image patches","volume":"34","author":"jing Zong","year":"2017","journal-title":"Biomed. Signal Process. Control."},{"issue":"3","key":"10.1016\/j.bspc.2026.110075_b31","doi-asserted-by":"crossref","first-page":"485","DOI":"10.1109\/LSP.2019.2895749","article-title":"Medical image fusion via convolutional sparsity based morphological component analysis","volume":"26","author":"Liu","year":"2019","journal-title":"IEEE Signal Process. Lett."},{"key":"10.1016\/j.bspc.2026.110075_b32","doi-asserted-by":"crossref","first-page":"1447","DOI":"10.1049\/iet-ipr.2018.6556","article-title":"Multi-sensor medical image fusion using pyramid-based DWT: a multi-resolution approach","volume":"13","author":"Nair","year":"2019","journal-title":"IET Image Process."},{"issue":"1","key":"10.1016\/j.bspc.2026.110075_b33","doi-asserted-by":"crossref","DOI":"10.1155\/2010\/579341","article-title":"Medical image fusion via an effective wavelet-based approach","volume":"2010","author":"Yang","year":"2010","journal-title":"EURASIP J. Adv. Signal Process."},{"issue":"10","key":"10.1016\/j.bspc.2026.110075_b34","doi-asserted-by":"crossref","first-page":"3735","DOI":"10.1109\/JSEN.2016.2533864","article-title":"Multimodal sensor medical image fusion based on type-2 fuzzy logic in NSCT domain","volume":"16","author":"Yang","year":"2016","journal-title":"IEEE Sensors J."},{"issue":"1","key":"10.1016\/j.bspc.2026.110075_b35","doi-asserted-by":"crossref","DOI":"10.1155\/2020\/6265708","article-title":"Brain medical image fusion based on dual-branch CNNs in NSST domain","volume":"2020","author":"Ding","year":"2020","journal-title":"BioMed Res. Int."},{"key":"10.1016\/j.bspc.2026.110075_b36","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2020.103823","article-title":"Multi-modal medical image fusion by Laplacian pyramid and adaptive sparse representation","volume":"123","author":"Wang","year":"2020","journal-title":"Comput. Biol. Med."},{"key":"10.1016\/j.bspc.2026.110075_b37","doi-asserted-by":"crossref","first-page":"302","DOI":"10.1016\/j.ins.2021.04.052","article-title":"Multimodal medical image fusion based on joint bilateral filter and local gradient energy","volume":"569","author":"Li","year":"2021","journal-title":"Inform. Sci."},{"issue":"9","key":"10.1016\/j.bspc.2026.110075_b38","doi-asserted-by":"crossref","first-page":"4630","DOI":"10.1109\/TCSVT.2023.3245607","article-title":"Laplacian pyramid fusion network with hierarchical guidance for infrared and visible image fusion","volume":"33","author":"Yao","year":"2023","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.bspc.2026.110075_b39","doi-asserted-by":"crossref","first-page":"147","DOI":"10.1016\/j.inffus.2014.09.004","article-title":"A general framework for image fusion based on multi-scale transform and sparse representation","volume":"24","author":"Liu","year":"2015","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.bspc.2026.110075_b40","doi-asserted-by":"crossref","first-page":"5134","DOI":"10.1109\/TIP.2022.3193288","article-title":"MATR: Multimodal medical image fusion via multiscale adaptive transformer","volume":"31","author":"Tang","year":"2022","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.bspc.2026.110075_b41","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102666","article-title":"MMIF-INet: Multimodal medical image fusion by invertible network","volume":"114","author":"He","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.bspc.2026.110075_b42","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2025.114053","article-title":"XKanFuse: A novel cross-modal fusion method based on Kolmogorov-Arnold network for multi-modal medical image fusion","volume":"326","author":"Wei","year":"2025","journal-title":"Knowl.-Based Syst."},{"issue":"11","key":"10.1016\/j.bspc.2026.110075_b43","doi-asserted-by":"crossref","first-page":"16248","DOI":"10.1109\/TNNLS.2023.3293274","article-title":"GeSeNet: A general semantic-guided network with couple mask ensemble for medical image fusion","volume":"35","author":"Li","year":"2024","journal-title":"IEEE Trans. Neural Networks Learn. Syst."},{"issue":"1","key":"10.1016\/j.bspc.2026.110075_b44","doi-asserted-by":"crossref","first-page":"79","DOI":"10.1162\/neco.1991.3.1.79","article-title":"Adaptive mixtures of local experts","volume":"3","author":"Jacobs","year":"1991","journal-title":"Neural Comput."},{"key":"10.1016\/j.bspc.2026.110075_b45","first-page":"8583","article-title":"Scaling vision with sparse mixture of experts","volume":"34","author":"Riquelme","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.bspc.2026.110075_b46","first-page":"9564","article-title":"Multimodal contrastive learning with limoe: the language-image mixture of experts","volume":"35","author":"Mustafa","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.bspc.2026.110075_b47","first-page":"2664","article-title":"Uni-perceiver-moe: Learning sparse generalist models with conditional moes","volume":"35","author":"Zhu","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.bspc.2026.110075_b48","first-page":"32897","article-title":"Vlmo: Unified vision-language pre-training with mixture-of-modality-experts","volume":"35","author":"Bao","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.bspc.2026.110075_b49","doi-asserted-by":"crossref","first-page":"23049","DOI":"10.52202\/068431-1675","article-title":"Towards understanding the mixture-of-experts layer in deep learning","volume":"35","author":"Chen","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.bspc.2026.110075_b50","doi-asserted-by":"crossref","unstructured":"K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770\u2013778.","DOI":"10.1109\/CVPR.2016.90"},{"issue":"120","key":"10.1016\/j.bspc.2026.110075_b51","first-page":"1","article-title":"Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity","volume":"23","author":"Fedus","year":"2022","journal-title":"J. Mach. Learn. Res."},{"issue":"4","key":"10.1016\/j.bspc.2026.110075_b52","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image quality assessment: from error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"issue":"10","key":"10.1016\/j.bspc.2026.110075_b53","doi-asserted-by":"crossref","first-page":"2761","DOI":"10.1007\/s11263-021-01501-8","article-title":"SDNet: A versatile squeeze-and-decomposition network for real-time image fusion","volume":"129","author":"Zhang","year":"2021","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.bspc.2026.110075_b54","first-page":"4725","article-title":"Bsafusion: A bidirectional stepwise feature alignment network for unaligned medical image fusion","volume":"vol. 39","author":"Li","year":"2025"},{"issue":"7","key":"10.1016\/j.bspc.2026.110075_b55","doi-asserted-by":"crossref","first-page":"971","DOI":"10.1016\/j.imavis.2007.10.012","article-title":"Multifocus image fusion using region segmentation and spatial frequency","volume":"26","author":"Li","year":"2008","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.bspc.2026.110075_b56","doi-asserted-by":"crossref","unstructured":"Z. Zhao, H. Bai, J. Zhang, Y. Zhang, S. Xu, Z. Lin, R. Timofte, L. Van Gool, Cddfuse: Correlation-driven dual-branch feature decomposition for multi-modality image fusion, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 5906\u20135916.","DOI":"10.1109\/CVPR52729.2023.00572"},{"issue":"12","key":"10.1016\/j.bspc.2026.110075_b57","doi-asserted-by":"crossref","first-page":"1890","DOI":"10.1016\/j.aeue.2015.09.004","article-title":"A new image quality metric for image fusion: The sum of the correlations of differences","volume":"69","author":"Aslantas","year":"2015","journal-title":"AEU - Int. J. Electron. Commun."},{"key":"10.1016\/j.bspc.2026.110075_b58","doi-asserted-by":"crossref","first-page":"313","DOI":"10.1049\/el:20020212","article-title":"Information measure for performance of image fusion","volume":"38","author":"Qu","year":"2002","journal-title":"Electron. Lett."},{"key":"10.1016\/j.bspc.2026.110075_b59","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2024.106561","article-title":"AMMNet: A multimodal medical image fusion method based on an attention mechanism and MobileNetV3","volume":"96","author":"Di","year":"2024","journal-title":"Biomed. Signal Process. Control."},{"key":"10.1016\/j.bspc.2026.110075_b60","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1007\/s11042-023-16892-8","article-title":"On the use of UDWT and fuzzy sets for medical image fusion","volume":"83","author":"Tirupal","year":"2023","journal-title":"Multimedia Tools Appl."},{"key":"10.1016\/j.bspc.2026.110075_b61","article-title":"Multimodal medical image fusion techniques \u2013 A review","volume":"15","author":"Tirupal","year":"2020","journal-title":"Curr. Signal Transduct. Ther."},{"key":"10.1016\/j.bspc.2026.110075_b62","doi-asserted-by":"crossref","first-page":"584","DOI":"10.1109\/TCI.2021.3083965","article-title":"Green fluorescent protein and phase contrast image fusion via detail preserving cross network","volume":"7","author":"Tang","year":"2021","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"10.1016\/j.bspc.2026.110075_b63","doi-asserted-by":"crossref","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","article-title":"IFCNN: A general image fusion framework based on convolutional neural network","volume":"54","author":"Zhang","year":"2020","journal-title":"Inf. Fusion"},{"issue":"7","key":"10.1016\/j.bspc.2026.110075_b64","doi-asserted-by":"crossref","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","article-title":"SwinFusion: Cross-domain long-range learning for general image fusion via swin transformer","volume":"9","author":"Ma","year":"2022","journal-title":"IEEE\/CAA J. Autom. Sin."},{"issue":"7","key":"10.1016\/j.bspc.2026.110075_b65","doi-asserted-by":"crossref","first-page":"3159","DOI":"10.1109\/TCSVT.2023.3234340","article-title":"DATFuse: Infrared and visible image fusion via dual attention transformer","volume":"33","author":"Tang","year":"2023","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.bspc.2026.110075_b66","doi-asserted-by":"crossref","first-page":"5413","DOI":"10.1109\/TMM.2022.3192661","article-title":"YDTR: Infrared and visible image fusion via Y-shape dynamic transformer","volume":"25","author":"Tang","year":"2023","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.bspc.2026.110075_b67","doi-asserted-by":"crossref","first-page":"111","DOI":"10.1016\/j.inffus.2021.02.005","article-title":"Benchmarking and comparing multi-exposure image fusion algorithms","author":"Zhang","year":"2021","journal-title":"Inf. Fusion"},{"issue":"1","key":"10.1016\/j.bspc.2026.110075_b68","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1109\/TFUZZ.2024.3407739","article-title":"Fcdnet: Fuzzy cognition-based dynamic fusion network for multimodal sentiment analysis","volume":"33","author":"Liu","year":"2025","journal-title":"IEEE Trans. Fuzzy Syst."},{"key":"10.1016\/j.bspc.2026.110075_b69","doi-asserted-by":"crossref","first-page":"679","DOI":"10.1016\/j.ins.2022.11.076","article-title":"Multi-modal fusion network with complementarity and importance for emotion recognition","volume":"619","author":"Liu","year":"2023","journal-title":"Inform. Sci."}],"container-title":["Biomedical Signal Processing and Control"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1746809426006294?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1746809426006294?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T04:48:48Z","timestamp":1776142128000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1746809426006294"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,7]]},"references-count":69,"alternative-id":["S1746809426006294"],"URL":"https:\/\/doi.org\/10.1016\/j.bspc.2026.110075","relation":{},"ISSN":["1746-8094"],"issn-type":[{"value":"1746-8094","type":"print"}],"subject":[],"published":{"date-parts":[[2026,7]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"A multi-scale expert gating network with cross-modal attention for multimodal medical image fusion","name":"articletitle","label":"Article Title"},{"value":"Biomedical Signal Processing and Control","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.bspc.2026.110075","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"110075"}}