{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,28]],"date-time":"2025-10-28T15:15:53Z","timestamp":1761664553386,"version":"3.37.3"},"reference-count":29,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2023,4,10]],"date-time":"2023-04-10T00:00:00Z","timestamp":1681084800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,4,10]],"date-time":"2023-04-10T00:00:00Z","timestamp":1681084800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100017700","name":"Henan Provincial Science and Technology Research Project","doi-asserted-by":"publisher","award":["No.222102210309"],"award-info":[{"award-number":["No.222102210309"]}],"id":[{"id":"10.13039\/501100017700","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100014718","name":"Innovative Research Group Project of the National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["Nos. 62106067","62106068"],"award-info":[{"award-number":["Nos. 62106067","62106068"]}],"id":[{"id":"10.13039\/100014718","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100006407","name":"Natural Science Foundation of Henan Province","doi-asserted-by":"publisher","award":["No.21A520010"],"award-info":[{"award-number":["No.21A520010"]}],"id":[{"id":"10.13039\/501100006407","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100010104","name":"Zhengzhou Municipal Science and Technology Bureau","doi-asserted-by":"publisher","award":["No. 21ZZXTCX21"],"award-info":[{"award-number":["No. 21ZZXTCX21"]}],"id":[{"id":"10.13039\/100010104","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100018607","name":"Henan Province Science and Technology Innovation Talent Program","doi-asserted-by":"publisher","award":["2021ZKCJ14"],"award-info":[{"award-number":["2021ZKCJ14"]}],"id":[{"id":"10.13039\/501100018607","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Digit Imaging"],"DOI":"10.1007\/s10278-023-00810-3","type":"journal-article","created":{"date-parts":[[2023,4,10]],"date-time":"2023-04-10T19:02:21Z","timestamp":1681153341000},"page":"1851-1863","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["RTFusion: A Multimodal Fusion Network with Significant Information Enhancement"],"prefix":"10.1007","volume":"36","author":[{"given":"Chao","family":"Fan","sequence":"first","affiliation":[]},{"given":"Zhixiang","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Xiao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Zhihui","family":"Xuan","sequence":"additional","affiliation":[]},{"given":"Zhentong","family":"Zhu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,10]]},"reference":[{"key":"810_CR1","unstructured":"L S Ambati, O F El-Gayar, N Nawar. Design principles for multiple sclerosis mobile self-management applications: a patient-centric perspective[C]\/\/ AMCIS 2021. 2021."},{"key":"810_CR2","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","volume":"76","author":"H Zhang","year":"2021","unstructured":"H. Zhang, H. Xu, X. Tian, et al. Image fusion meets deep learning: a survey and perspective. Information Fusion, vol. 76, pp. 323\u2013336, 2021.","journal-title":"Information Fusion"},{"key":"810_CR3","doi-asserted-by":"crossref","unstructured":"Qiang Zhang, Yi Liu, Rick S. Blum, et al. Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: a review[J]. Information Fusion, 2018.","DOI":"10.1016\/j.inffus.2017.05.006"},{"key":"810_CR4","doi-asserted-by":"crossref","unstructured":"Y. Liu, X. Chen, J. Cheng, et al. A medical image fusion method based on convolutional neural networks. Proceedings of the International Conference on Information Fusion, 2017, pp. 1\u20137.","DOI":"10.23919\/ICIF.2017.8009769"},{"key":"810_CR5","doi-asserted-by":"crossref","unstructured":"S Li, X Kang, L Fang, et al. Pixel-level image fusion: a survey of the state of the art[J]. Information Fusion, 2017, 33.","DOI":"10.1016\/j.inffus.2016.05.004"},{"key":"810_CR6","unstructured":"P J Burt, K Hanna, R J Kolczynski. Enhanced image capture through fusion[C]\/\/ 1993 (4th) International Conference on Computer Vision. IEEE, 1993."},{"issue":"23","key":"810_CR7","doi-asserted-by":"publisher","first-page":"4184","DOI":"10.1016\/j.ijleo.2015.08.118","volume":"126","author":"W Zhi-She","year":"2015","unstructured":"Zhi-She W, Feng-Bao Y, Zhi-Hao P, et al. Multi-sensor image enhanced fusion algorithm based on NSST and top-hat transformation[J]. Optik - International Journal for Light and Electron Optics, 2015, 126(23):4184-4190.","journal-title":"Optik - International Journal for Light and Electron Optics"},{"key":"810_CR8","unstructured":"A Zw, B Jx, A Xj, et al. Infrared and visible image fusion via hybrid decomposition of NSCT and morphological sequential toggle operator[J]. Optik, 201."},{"issue":"2","key":"810_CR9","doi-asserted-by":"publisher","first-page":"220","DOI":"10.1109\/LSP.2014.2354534","volume":"22","author":"L Cao","year":"2014","unstructured":"L Cao, L Jin, H Tao, et al. Multi-focus image fusion based on spatial frequency in discrete cosine transform domain[J]. IEEE Signal Processing Letters, 2014, 22(2):220-224.","journal-title":"IEEE Signal Processing Letters"},{"key":"810_CR10","doi-asserted-by":"crossref","unstructured":"Y. Liu, X. Chen, R. K. et al, Medical image fusion via convolutional sparsity based morphological component analysis. IEEE Signal Processing Letters, vol. 26, no. 3, pp. 485\u2013489, 2019.","DOI":"10.1109\/LSP.2019.2895749"},{"key":"810_CR11","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1016\/j.inffus.2014.05.004","volume":"23","author":"Y Liu","year":"2015","unstructured":"Y. Liu, S. Liu, Z. Wang. Multi-focus image fusion with dense sift. Information Fusion, vol. 23, pp. 139\u2013155, 2015.","journal-title":"Information Fusion"},{"issue":"5","key":"810_CR12","doi-asserted-by":"publisher","first-page":"2519","DOI":"10.1109\/TIP.2017.2671921","volume":"26","author":"K Ma","year":"2017","unstructured":"K. Ma, H. Li, H. Yong, et al. Robust multi-exposure image fusion: a structural patch decomposition approach. IEEE Transactions on Image Processing, vol. 26, no. 5, pp. 2519\u20132532, 2017.","journal-title":"IEEE Transactions on Image Processing"},{"key":"810_CR13","doi-asserted-by":"publisher","unstructured":"Yu Zhang, Yu Liu, Peng Sun, et al. IFCNN: a general image fusion framework based on convolutional neural network. Information Fusion. 54. https:\/\/doi.org\/10.1016\/j.inffus.2019.07.011.","DOI":"10.1016\/j.inffus.2019.07.011"},{"key":"810_CR14","doi-asserted-by":"crossref","unstructured":"K. Ram Prabhakar, V. Sai Srikar, R. Venkatesh Babu. DeepFuse: a deep unsupervised approach for exposure fusion with extreme exposure image pairs. Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 4714\u20134722.","DOI":"10.1109\/ICCV.2017.505"},{"key":"810_CR15","doi-asserted-by":"crossref","unstructured":"H Xu, J Ma, Z Le, et al. FusionDN: a unified densely connected network for image fusion[C]\/\/ AAAI. 2020:12484\u201312491.","DOI":"10.1609\/aaai.v34i07.6936"},{"key":"810_CR16","doi-asserted-by":"publisher","unstructured":"H. Xu, J. Ma, J. Jiang, et al. U2Fusion: a unified unsupervised image fusion network. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 1, pp. 502\u2013518, 1 Jan. 2022, https:\/\/doi.org\/10.1109\/TPAMI.2020.3012548.","DOI":"10.1109\/TPAMI.2020.3012548"},{"key":"810_CR17","doi-asserted-by":"crossref","unstructured":"Han Xu, Jiayi Ma. EMFusion: an unsupervised enhanced medical image fusion network, Information Fusion, Volume 76,2021, Pages 177\u2013186, ISSN 1566\u20132535,","DOI":"10.1016\/j.inffus.2021.06.001"},{"key":"810_CR18","first-page":"6","volume":"3","author":"A Dosovitskiy","year":"2020","unstructured":"A. Dosovitskiy, L. Beyer, A. Kolesnikov, et al. An image is worth 16x16 words: transformers for image recognition at scale. International Conference on Learning Representations, 2020. 3, 6","journal-title":"International Conference on Learning Representations"},{"issue":"7","key":"810_CR19","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"Jiayi Ma","year":"2022","unstructured":"Jiayi Ma, Linfeng Tang, Fan Fan, et al. SwinFusion: cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA J. Autom. Sinica, vol. 9, no. 7, pp. 1200-1217, Jul. 2022. https:\/\/doi.org\/10.1109\/JAS.2022.105686","journal-title":"IEEE\/CAA J. Autom. Sinica"},{"key":"810_CR20","unstructured":"T. Xiao, P. Dollar, M. Singh, et al. Early convolutions help transformers see better. Advances in Neural Information Processing Systems, 2021, pp. 30 392\u201330 400."},{"key":"810_CR21","doi-asserted-by":"crossref","unstructured":"Tang, Wei, et al. \"MATR: multi-modal medical image fusion via multiscale adaptive transformer.\" IEEE Transactions on Image Processing 31 (2022): 5134\u20135149.","DOI":"10.1109\/TIP.2022.3193288"},{"key":"810_CR22","doi-asserted-by":"crossref","unstructured":"Zhou, Quan, et al. \"Multi-modal medical image fusion based on densely-connected high-resolution CNN and hybrid transformer.\" Neural Computing and Applications 34.24 (2022): 21741\u201321761.","DOI":"10.1007\/s00521-022-07635-1"},{"key":"810_CR23","doi-asserted-by":"crossref","unstructured":"Li, Weisheng, et al. \"DFENet: A dual-branch feature enhanced network integrating transformers and convolutional feature learning for multi-modal medical image fusion.\" Biomedical Signal Processing and Control 80 (2023): 104402.","DOI":"10.1016\/j.bspc.2022.104402"},{"key":"810_CR24","unstructured":"Michael Hassid, Hao Peng, Daniel Rotem, et al. How much does attention actually attend? Questioning the Importance of Attention in Pretrained Transformers. arXiv:2211.03495"},{"key":"810_CR25","doi-asserted-by":"publisher","unstructured":"Laung-Terng Wang, N. E. Hoover, E. H. Porter, et al. SSIM: a software levelized compiled-code simulator. 24th ACM\/IEEE Design Automation Conference, 1987, pp. 2\u20138, https:\/\/doi.org\/10.1109\/DAC.1987.203214.","DOI":"10.1109\/DAC.1987.203214"},{"key":"810_CR26","doi-asserted-by":"crossref","unstructured":"CS Xydeas, VS Petrovic. Objective pixel-level image fusion performance measure[C]\/\/ Aerosense. International Society for Optics and Photonics, 2000.","DOI":"10.1117\/12.381668"},{"key":"810_CR27","doi-asserted-by":"crossref","unstructured":"G. Qu, D. Zhang, P. Yan. Information measure for performance of image fusion. Electronics letters, vol. 38, no. 7, pp. 313\u2013315, 2002. 8.","DOI":"10.1049\/el:20020212"},{"key":"810_CR28","doi-asserted-by":"crossref","unstructured":"Mohammad Bagher Akbari Haghighat, Ali Aghagolzadeh, Hadi Seyedarabi. A non-reference image fusion metric based on mutual information of image features. Computers & Electrical Engineering, Volume 37, Issue 5, 2011, Pages 744\u2013756, ISSN 0045\u20137906.","DOI":"10.1016\/j.compeleceng.2011.07.012"},{"key":"810_CR29","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"L Tang, J Yuan, J Ma. Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network[J]. Information Fusion, 2022, 82:28-42.","journal-title":"Information Fusion"}],"container-title":["Journal of Digital Imaging"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10278-023-00810-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10278-023-00810-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10278-023-00810-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,8,7]],"date-time":"2023-08-07T17:10:02Z","timestamp":1691428202000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10278-023-00810-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,10]]},"references-count":29,"journal-issue":{"issue":"4","published-online":{"date-parts":[[2023,8]]}},"alternative-id":["810"],"URL":"https:\/\/doi.org\/10.1007\/s10278-023-00810-3","relation":{},"ISSN":["1618-727X"],"issn-type":[{"type":"electronic","value":"1618-727X"}],"subject":[],"published":{"date-parts":[[2023,4,10]]},"assertion":[{"value":"14 December 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 March 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 March 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 April 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing of Interests"}}]}}