{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:16:21Z","timestamp":1775578581525,"version":"3.50.1"},"reference-count":49,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T00:00:00Z","timestamp":1755734400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T00:00:00Z","timestamp":1755734400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/100014718","name":"Innovative Research Group Project of the National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61671470"],"award-info":[{"award-number":["61671470"]}],"id":[{"id":"10.13039\/100014718","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100014718","name":"Innovative Research Group Project of the National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61671470"],"award-info":[{"award-number":["61671470"]}],"id":[{"id":"10.13039\/100014718","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s40747-025-02046-w","type":"journal-article","created":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T06:32:46Z","timestamp":1755757966000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["DBSQFusion: a multimodal image fusion method based on dual-channel attention"],"prefix":"10.1007","volume":"11","author":[{"ORCID":"https:\/\/orcid.org\/0009-0000-9075-3505","authenticated-orcid":false,"given":"Shaodong","family":"Liu","sequence":"first","affiliation":[]},{"given":"Faming","family":"Shao","sequence":"additional","affiliation":[]},{"given":"Xiaohui","family":"He","sequence":"additional","affiliation":[]},{"given":"Jinhong","family":"Xue","sequence":"additional","affiliation":[]},{"given":"Heng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Qing","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,21]]},"reference":[{"key":"2046_CR1","doi-asserted-by":"publisher","first-page":"102536","DOI":"10.1016\/j.inffus.2024.102536","volume":"112","author":"J Duan","year":"2024","unstructured":"Duan J, Xiong J, Li Y, Ding W (2024) Deep learning based multimodal biomedical data fusion: an overview and comparative review. Inf Fusion 112:102536. https:\/\/doi.org\/10.1016\/j.inffus.2024.102536","journal-title":"Inf Fusion"},{"key":"2046_CR2","doi-asserted-by":"publisher","first-page":"106959","DOI":"10.1016\/j.compbiomed.2023.106959","volume":"160","author":"T Zhou","year":"2023","unstructured":"Zhou T, Cheng Q, Lu H, Li Q, Zhang X, Qiu S (2023) Deep learning methods for medical image fusion: a review. Comput Biol Med 160:106959. https:\/\/doi.org\/10.1016\/j.compbiomed.2023.106959","journal-title":"Comput Biol Med"},{"key":"2046_CR3","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1016\/j.inffus.2022.09.008","volume":"90","author":"Z Wang","year":"2023","unstructured":"Wang Z, Ma Y, Zhang Y (2023) Review of pixel-level remote sensing image fusion based on deep learning. Inf Fusion 90:36\u201358. https:\/\/doi.org\/10.1016\/j.inffus.2022.09.008","journal-title":"Inf Fusion"},{"key":"2046_CR4","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"J Ma","year":"2019","unstructured":"Ma J, Ma Y, Li C (2019) Infrared and visible image fusion methods and applications: a survey. Inf Fusion 45:153\u2013178. https:\/\/doi.org\/10.1016\/j.inffus.2018.02.004","journal-title":"Inf Fusion"},{"key":"2046_CR5","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.2003.09210","author":"Z Zhao","year":"2020","unstructured":"Zhao Z, Xu S, Zhang C, Liu J, Li P, Zhang J (2020) DIDFuse: Deep Image Decomposition for Infrared and Visible Image Fusion. Arxiv Preprint. https:\/\/doi.org\/10.48550\/ARXIV.2003.09210","journal-title":"Arxiv Preprint"},{"key":"2046_CR6","doi-asserted-by":"publisher","first-page":"1771","DOI":"10.1016\/j.patcog.2006.11.010","volume":"40","author":"J Han","year":"2007","unstructured":"Han J, Bhanu B (2007) Fusion of color and infrared video for moving human detection. Pattern Recogn 40:1771\u20131784. https:\/\/doi.org\/10.1016\/j.patcog.2006.11.010","journal-title":"Pattern Recogn"},{"key":"2046_CR7","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li H, Wu X-J, Kittler J (2021) RFN-nest: an end-to-end residual fusion network for infrared and visible images. Inf Fusion 73:72\u201386. https:\/\/doi.org\/10.1016\/j.inffus.2021.02.023","journal-title":"Inf Fusion"},{"key":"2046_CR8","doi-asserted-by":"publisher","first-page":"880","DOI":"10.1016\/j.patcog.2007.06.022","volume":"41","author":"R Singh","year":"2008","unstructured":"Singh R, Vatsa M, Noore A (2008) Integrated multilevel image fusion and match score fusion of visible and infrared face images for robust face recognition. Pattern Recogn 41:880\u2013893. https:\/\/doi.org\/10.1016\/j.patcog.2007.06.022","journal-title":"Pattern Recogn"},{"key":"2046_CR9","doi-asserted-by":"publisher","first-page":"9909","DOI":"10.1109\/TVT.2023.3259999","volume":"72","author":"Y An","year":"2023","unstructured":"An Y, Wu J, Cui Y, Hu H (2023) Multi-object tracking based on a novel feature image with multi-modal information. IEEE Trans Veh Technol 72:9909\u20139921. https:\/\/doi.org\/10.1109\/TVT.2023.3259999","journal-title":"IEEE Trans Veh Technol"},{"key":"2046_CR10","doi-asserted-by":"publisher","first-page":"117228","DOI":"10.1016\/j.image.2024.117228","volume":"130","author":"Y Luo","year":"2025","unstructured":"Luo Y, He K, Xu D, Shi H, Yin W (2025) Infrared and visible image fusion based on hybrid multi-scale decomposition and adaptive contrast enhancement. Signal Process Image Commun 130:117228. https:\/\/doi.org\/10.1016\/j.image.2024.117228","journal-title":"Signal Process Image Commun"},{"key":"2046_CR11","doi-asserted-by":"publisher","first-page":"102871","DOI":"10.1016\/j.inffus.2024.102871","volume":"117","author":"T Wang","year":"2025","unstructured":"Wang T, Yu Z, Fang J, Xie J, Yang F, Zhang H, Zhang L, Du M, Li L, Ning X (2025) Multidimensional fusion of frequency and spatial domain information for enhanced camouflaged object detection. Inf Fusion 117:102871. https:\/\/doi.org\/10.1016\/j.inffus.2024.102871","journal-title":"Inf Fusion"},{"key":"2046_CR12","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2023.3267495","volume":"61","author":"S Wang","year":"2023","unstructured":"Wang S, Cai Z, Yuan J (2023) Automatic SAR ship detection based on multifeature fusion network in spatial and frequency domains. IEEE Trans Geosci Remote Sens 61:1\u201311. https:\/\/doi.org\/10.1109\/TGRS.2023.3267495","journal-title":"IEEE Trans Geosci Remote Sens"},{"key":"2046_CR13","doi-asserted-by":"publisher","first-page":"119909","DOI":"10.1016\/j.eswa.2023.119909","volume":"223","author":"L Qu","year":"2023","unstructured":"Qu L, Yin S, Liu S, Liu X, Wang M, Song Z (2023) AIM-MEF: multi-exposure image fusion based on adaptive information mining in both spatial and frequency domains. Expert Syst Appl 223:119909. https:\/\/doi.org\/10.1016\/j.eswa.2023.119909","journal-title":"Expert Syst Appl"},{"key":"2046_CR14","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1016\/j.inffus.2021.04.005","volume":"75","author":"Y Mo","year":"2021","unstructured":"Mo Y, Kang X, Duan P, Sun B, Li S (2021) Attribute filter based infrared and visible image fusion. Inf Fusion 75:41\u201354. https:\/\/doi.org\/10.1016\/j.inffus.2021.04.005","journal-title":"Inf Fusion"},{"key":"2046_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2023.3328710","volume":"72","author":"L Chang","year":"2023","unstructured":"Chang L, Huang Y, Li Q, Su W, Zhang Y (2023) L2 net: infrared and visible image fusion using lightweight large kernel convolution network. IEEE Trans Instrum Meas 72:1\u201313. https:\/\/doi.org\/10.1109\/TIM.2023.3328710","journal-title":"IEEE Trans Instrum Meas"},{"key":"2046_CR16","doi-asserted-by":"publisher","first-page":"104741","DOI":"10.1016\/j.dsp.2024.104741","volume":"155","author":"H Xu","year":"2024","unstructured":"Xu H, Nie R, Cao J, Tan M, Ding Z (2024) MADmfuse: a multi-attribute diffusion model to fuse infrared and visible images. Digit Signal Process 155:104741. https:\/\/doi.org\/10.1016\/j.dsp.2024.104741","journal-title":"Digit Signal Process"},{"key":"2046_CR17","doi-asserted-by":"publisher","first-page":"1041","DOI":"10.1016\/j.asoc.2011.11.020","volume":"12","author":"J Saeedi","year":"2012","unstructured":"Saeedi J, Faez K (2012) Infrared and Visible Image Fusion Using Fuzzy Logic and Population-Based Optimization. Appl Soft Comput 12:1041\u20131054. https:\/\/doi.org\/10.1016\/j.asoc.2011.11.020","journal-title":"Appl Soft Comput"},{"key":"2046_CR18","doi-asserted-by":"publisher","first-page":"28","DOI":"10.28991\/ESJ-2024-08-01-03","volume":"8","author":"BU Mahmud","year":"2024","unstructured":"Mahmud BU, Mamun AA, Hossen MJ, Hong GY, Jahan B (2024) Light-weight deep learning model for accelerating the classification of mango-leaf disease. Emerg Sci J 8:28\u201342","journal-title":"Emerg Sci J"},{"key":"2046_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2024.3450061","volume":"73","author":"X Hu","year":"2024","unstructured":"Hu X, Liu Y, Yang F (2024) PFCfuse: a poolformer and CNN fusion network for infrared-visible image fusion. IEEE Trans Instrum Meas 73:1\u201314. https:\/\/doi.org\/10.1109\/TIM.2024.3450061","journal-title":"IEEE Trans Instrum Meas"},{"key":"2046_CR20","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu Y, Chen X, Peng H, Wang Z (2017) Multi-focus image fusion with a deep convolutional neural network. Inf Fusion 36:191\u2013207. https:\/\/doi.org\/10.1016\/j.inffus.2016.12.001","journal-title":"Inf Fusion"},{"key":"2046_CR21","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2019","unstructured":"Li H, Wu X-J (2019) Densefuse: a fusion approach to infrared and visible images. IEEE Trans Image Process 28:2614\u20132623. https:\/\/doi.org\/10.1109\/TIP.2018.2887342","journal-title":"IEEE Trans Image Process"},{"key":"2046_CR22","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","volume":"69","author":"H Li","year":"2020","unstructured":"Li H, Wu X-J, Durrani T (2020) Nestfuse: an infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models. IEEE Trans Instrum Meas 69:9645\u20139656. https:\/\/doi.org\/10.1109\/TIM.2020.3005230","journal-title":"IEEE Trans Instrum Meas"},{"key":"2046_CR23","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma J, Yu W, Liang P, Li C, Jiang J (2019) FusionGAN: a generative adversarial network for infrared and visible image fusion. Inf Fusion 48:11\u201326. https:\/\/doi.org\/10.1016\/j.inffus.2018.09.004","journal-title":"Inf Fusion"},{"key":"2046_CR24","doi-asserted-by":"publisher","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","volume":"29","author":"J Ma","year":"2020","unstructured":"Ma J, Xu H, Jiang J, Mei X, Zhang X-P (2020) DDcGAN: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans Image Process 29:4980\u20134995. https:\/\/doi.org\/10.1109\/TIP.2020.2977573","journal-title":"IEEE Trans Image Process"},{"key":"2046_CR25","doi-asserted-by":"publisher","first-page":"121156","DOI":"10.1016\/j.eswa.2023.121156","volume":"235","author":"Z Duan","year":"2024","unstructured":"Duan Z, Luo X, Zhang T (2024) Combining transformers with CNN for multi-focus image fusion. Expert Syst Appl 235:121156. https:\/\/doi.org\/10.1016\/j.eswa.2023.121156","journal-title":"Expert Syst Appl"},{"key":"2046_CR26","doi-asserted-by":"publisher","first-page":"125427","DOI":"10.1016\/j.eswa.2024.125427","volume":"261","author":"J Li","year":"2025","unstructured":"Li J, Song H, Liu L, Li Y, Xia J, Huang Y, Fan J, Lin Y, Yang J (2025) Mixfuse: an iterative mix-attention transformer for multi-modal image fusion. Expert Syst Appl 261:125427. https:\/\/doi.org\/10.1016\/j.eswa.2024.125427","journal-title":"Expert Syst Appl"},{"key":"2046_CR27","doi-asserted-by":"publisher","first-page":"111476","DOI":"10.1016\/j.patcog.2025.111476","volume":"163","author":"Q Lu","year":"2025","unstructured":"Lu Q, Zhang H, Yin L (2025) Infrared and visible image fusion via dual encoder based on dense connection. Pattern Recogn 163:111476. https:\/\/doi.org\/10.1016\/j.patcog.2025.111476","journal-title":"Pattern Recogn"},{"key":"2046_CR28","doi-asserted-by":"publisher","first-page":"102931","DOI":"10.1016\/j.inffus.2025.102931","volume":"118","author":"X Wang","year":"2025","unstructured":"Wang X, Guan Z, Qian W, Cao J, Ma R, Bi C (2025) A degradation-aware guided fusion network for infrared and visible image. Inf Fusion 118:102931. https:\/\/doi.org\/10.1016\/j.inffus.2025.102931","journal-title":"Inf Fusion"},{"key":"2046_CR29","doi-asserted-by":"publisher","first-page":"108800","DOI":"10.1016\/j.optlaseng.2024.108800","volume":"186","author":"Z Pei","year":"2025","unstructured":"Pei Z, Lu J, Qian Y, Fan L, Wang H, Chen J (2025) A new method for fusing infrared and visible images in low-light environments based on visual perception and attention mechanism. Opt Lasers Eng 186:108800. https:\/\/doi.org\/10.1016\/j.optlaseng.2024.108800","journal-title":"Opt Lasers Eng"},{"key":"2046_CR30","doi-asserted-by":"publisher","first-page":"110223","DOI":"10.1016\/j.patcog.2023.110223","volume":"149","author":"X Yang","year":"2024","unstructured":"Yang X, Huo H, Li C, Liu X, Wang W, Wang C (2024) Semantic perceptive infrared and visible image fusion transformer. Pattern Recogn 149:110223. https:\/\/doi.org\/10.1016\/j.patcog.2023.110223","journal-title":"Pattern Recogn"},{"key":"2046_CR31","doi-asserted-by":"publisher","first-page":"20139","DOI":"10.1007\/s11042-022-14314-9","volume":"82","author":"H Liu","year":"2023","unstructured":"Liu H, Yan H (2023) An end-to-end multi-scale network based on autoencoder for infrared and visible image fusion. Multimed Tools Appl 82:20139\u201320156. https:\/\/doi.org\/10.1007\/s11042-022-14314-9","journal-title":"Multimed Tools Appl"},{"key":"2046_CR32","doi-asserted-by":"publisher","first-page":"109620","DOI":"10.1016\/j.sigpro.2024.109620","volume":"225","author":"B Zheng","year":"2024","unstructured":"Zheng B, Huo H, Liu X, Pang S, Li J (2024) Pedestrian detection-driven cascade network for infrared and visible image fusion. Signal Process 225:109620. https:\/\/doi.org\/10.1016\/j.sigpro.2024.109620","journal-title":"Signal Process"},{"key":"2046_CR33","doi-asserted-by":"publisher","first-page":"111098","DOI":"10.1016\/j.patcog.2024.111098","volume":"159","author":"X Chen","year":"2025","unstructured":"Chen X, Xu S, Hu S, Ma X (2025) ACFNet: an adaptive cross-fusion network for infrared and visible image fusion. Pattern Recogn 159:111098. https:\/\/doi.org\/10.1016\/j.patcog.2024.111098","journal-title":"Pattern Recogn"},{"key":"2046_CR34","doi-asserted-by":"publisher","first-page":"477","DOI":"10.1016\/j.inffus.2022.10.034","volume":"91","author":"L Tang","year":"2023","unstructured":"Tang L, Xiang X, Zhang H, Gong M, Ma J (2023) Divfusion: darkness-free infrared and visible image fusion. Inf Fusion 91:477\u2013493. https:\/\/doi.org\/10.1016\/j.inffus.2022.10.034","journal-title":"Inf Fusion"},{"key":"2046_CR35","doi-asserted-by":"publisher","unstructured":"Gulrajani I, Ahmed F, Arjovsky M, Dumoulin V, Courville A (2017) Improved Training of Wasserstein GANs. https:\/\/doi.org\/10.48550\/arXiv.1704.00028","DOI":"10.48550\/arXiv.1704.00028"},{"key":"2046_CR36","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.08500","author":"M Heusel","year":"2017","unstructured":"Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S (2017) GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. Adv Neural Inform Process Syst. https:\/\/doi.org\/10.48550\/ARXIV.1706.08500","journal-title":"Adv Neural Inform Process Syst."},{"key":"2046_CR37","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1901.10422","author":"D Zhang","year":"2018","unstructured":"Zhang D, Khoreva A (2018) PA-GAN: Improving GAN Training by Progressive Augmentation. arXivorg. https:\/\/doi.org\/10.48550\/arXiv.1901.10422","journal-title":"arXivorg"},{"key":"2046_CR38","doi-asserted-by":"publisher","first-page":"5413","DOI":"10.1109\/TMM.2022.3192661","volume":"25","author":"W Tang","year":"2023","unstructured":"Tang W, He F, Liu Y (2023) YDTR: infrared and visible image fusion via Y-shape dynamic transformer. IEEE Trans Multimedia 25:5413\u20135428. https:\/\/doi.org\/10.1109\/TMM.2022.3192661","journal-title":"IEEE Trans Multimedia"},{"key":"2046_CR39","doi-asserted-by":"publisher","first-page":"73","DOI":"10.1007\/s10278-015-9806-4","volume":"29","author":"P Ganasala","year":"2016","unstructured":"Ganasala P, Kumar V (2016) Feature-motivated simplified adaptive PCNN-based medical image fusion algorithm in NSST domain. J Digit Imaging 29:73\u201385. https:\/\/doi.org\/10.1007\/s10278-015-9806-4","journal-title":"J Digit Imaging"},{"key":"2046_CR40","doi-asserted-by":"publisher","unstructured":"Zhao X, Jia H, Pang Y,  Lv L, Tian F, Zhang L, Sun W, Lu H (2023) M2SNet: Multi-Scale in Multi-Scale Subtraction Network for Medical Image Segmentation. https:\/\/doi.org\/10.48550\/arXiv.2303.10894","DOI":"10.48550\/arXiv.2303.10894"},{"key":"2046_CR41","doi-asserted-by":"publisher","unstructured":"Jia X, Zhu C, Li M, Tang W, Liu S, Zhou W (2021) LLVIP: A Visible-Infrared Paired Dataset\nfor Low-Light Vision. https:\/\/doi.org\/10.48550\/arXiv.2108.10831","DOI":"10.48550\/arXiv.2108.10831"},{"key":"2046_CR42","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","volume":"83","author":"L Tang","year":"2022","unstructured":"Tang L, Yuan J, Zhang H, Jiang X, Ma J (2022) Piafusion: a progressive infrared and visible image fusion network based on illumination aware. Inf Fusion 83:79\u201392. https:\/\/doi.org\/10.1016\/j.inffus.2022.03.007","journal-title":"Inf Fusion"},{"key":"2046_CR43","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1016\/j.dib.2017.09.038","volume":"15","author":"A Toet","year":"2017","unstructured":"Toet A (2017) The TNO multiband image data collection. Data Brief 15:249\u2013251. https:\/\/doi.org\/10.1016\/j.dib.2017.09.038","journal-title":"Data Brief"},{"key":"2046_CR44","doi-asserted-by":"publisher","first-page":"3804","DOI":"10.3390\/rs16203804","volume":"16","author":"L Li","year":"2024","unstructured":"Li L, Shi Y, Lv M, Jia Z, Liu M, Zhao X, Zhang X, Ma H (2024) Infrared and visible image fusion via sparse representation and guided filtering in Laplacian pyramid domain. Remote Sens 16:3804. https:\/\/doi.org\/10.3390\/rs16203804","journal-title":"Remote Sens"},{"key":"2046_CR45","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2024.3457951","volume":"73","author":"G Lv","year":"2024","unstructured":"Lv G, Sima C, Gao Y, Dong A, Ma G, Cheng J (2024) Sigfusion: semantic information-guided infrared and visible image fusion. IEEE Trans Instrum Meas 73:1\u201318. https:\/\/doi.org\/10.1109\/TIM.2024.3457951","journal-title":"IEEE Trans Instrum Meas"},{"key":"2046_CR46","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/tpami.2020.3012548","volume":"44","author":"H Xu","year":"2022","unstructured":"Xu H, Ma J, Jiang J, Guo X, Ling H (2022) U2fusion: a unified unsupervised image fusion network. IEEE Trans Pattern Anal Mach Intell 44:502\u2013518. https:\/\/doi.org\/10.1109\/tpami.2020.3012548","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"2046_CR47","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2301.08072","author":"J Yue","year":"2023","unstructured":"Yue J, Fang L, Xia S, Deng Y, Ma J (2023) Dif-Fusion: Towards High Color Fidelity in Infrared and Visible Image Fusion with Diffusion Models. IEEE Trans Image Process. https:\/\/doi.org\/10.48550\/arXiv.2301.08072","journal-title":"IEEE Trans Image Process."},{"key":"2046_CR48","doi-asserted-by":"publisher","first-page":"1044","DOI":"10.3390\/rs17061044","volume":"17","author":"S Liu","year":"2025","unstructured":"Liu S, Shao F, Chu W, Dai J, Zhang H (2025) An improved YOLOv8-based lightweight attention mechanism for cross-scale feature fusion. Remote Sens 17:1044. https:\/\/doi.org\/10.3390\/rs17061044","journal-title":"Remote Sens"},{"key":"2046_CR49","doi-asserted-by":"publisher","first-page":"677","DOI":"10.28991\/HIJ-2024-05-03-09","volume":"5","author":"M Zhang","year":"2024","unstructured":"Zhang M (2024) An improved fire detection algorithm based on YOLOv8 integrated with DGIConv, FourBranchAttention and GSIoU. HighTech Innov J 5:677\u2013689. https:\/\/doi.org\/10.28991\/HIJ-2024-05-03-09","journal-title":"HighTech Innov J"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-02046-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-025-02046-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-025-02046-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,25]],"date-time":"2025-09-25T13:33:19Z","timestamp":1758807199000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-025-02046-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,21]]},"references-count":49,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["2046"],"URL":"https:\/\/doi.org\/10.1007\/s40747-025-02046-w","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,8,21]]},"assertion":[{"value":"18 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 July 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 August 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}],"article-number":"421"}}