{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T08:24:31Z","timestamp":1765355071569,"version":"3.37.3"},"reference-count":55,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,9,9]],"date-time":"2024-09-09T00:00:00Z","timestamp":1725840000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,9,9]],"date-time":"2024-09-09T00:00:00Z","timestamp":1725840000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Gansu Natural Science Foundation Programme","award":["24JRRA231"],"award-info":[{"award-number":["24JRRA231"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62061023"],"award-info":[{"award-number":["62061023"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Distinguished Young Scholars of Gansu Province of China","award":["21JR7RA345"],"award-info":[{"award-number":["21JR7RA345"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,10]]},"DOI":"10.1007\/s00530-024-01473-y","type":"journal-article","created":{"date-parts":[[2024,9,9]],"date-time":"2024-09-09T12:02:10Z","timestamp":1725883330000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["C2IENet: Multi-branch medical image fusion based on contrastive constraint features and information exchange"],"prefix":"10.1007","volume":"30","author":[{"given":"Jing","family":"Di","sequence":"first","affiliation":[]},{"given":"Chan","family":"Liang","sequence":"additional","affiliation":[]},{"given":"Li","family":"Ren","sequence":"additional","affiliation":[]},{"given":"Wenqing","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Jizhao","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jing","family":"Lian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,9]]},"reference":[{"key":"1473_CR1","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","volume":"76","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Xu, H., Tian, X., Jiang, J., Ma, J.: Image fusion meets deep learning: a survey and perspective. Inform. Fusion 76, 323\u2013336 (2021). https:\/\/doi.org\/10.1016\/j.inffus.2021.06.008","journal-title":"Inform. Fusion"},{"key":"1473_CR2","doi-asserted-by":"publisher","first-page":"1261","DOI":"10.1109\/TIP.2020.3043125","volume":"30","author":"R Liu","year":"2020","unstructured":"Liu, R., Liu, J., Jiang, Z., Fan, X., Luo, Z.: A bilevel integrated model with data-driven layer ensemble for multi-modality image fusion. IEEE Trans. Image Process. 30, 1261\u20131274 (2020). https:\/\/doi.org\/10.1109\/TIP.2020.3043125","journal-title":"IEEE Trans. Image Process."},{"key":"1473_CR3","doi-asserted-by":"publisher","first-page":"103","DOI":"10.1016\/j.inffus.2022.10.010","volume":"91","author":"H Zhang","year":"2023","unstructured":"Zhang, H., Wang, H., Tian, X., Ma, J.: P2Sharpen: a progressive pansharpening network with deep spectral transformation. Inform. Fusion 91, 103\u2013122 (2023). https:\/\/doi.org\/10.1016\/j.inffus.2022.10.010","journal-title":"Inform. Fusion"},{"issue":"12","key":"1473_CR4","doi-asserted-by":"publisher","first-page":"7583","DOI":"10.1002\/mp.15882","volume":"49","author":"S Liu","year":"2022","unstructured":"Liu, S., Tang, X., Cai, T., Zhang, Y., Wang, C.: COVID-19 CT image segmentation based on improved Res2Net. Med. Phys. 49(12), 7583\u20137595 (2022). https:\/\/doi.org\/10.1002\/mp.15882","journal-title":"Med. Phys."},{"key":"1473_CR5","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1016\/j.neucom.2022.01.059","volume":"480","author":"G Wang","year":"2022","unstructured":"Wang, G., Li, W., Gao, X., Xiao, B., Du, J.: Multimodal medical image fusion based on multichannel coupled neural P systems and max-cloud models in spectral total variation domain. Neurocomputing 480, 61\u201375 (2022). https:\/\/doi.org\/10.1016\/j.neucom.2022.01.059","journal-title":"Neurocomputing"},{"key":"1473_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2022.108542","volume":"118","author":"Z Chao","year":"2022","unstructured":"Chao, Z., Duan, X., Jia, S., Guo, X., Liu, H., Jia, F.: Medical image fusion via discrete stationary wavelet transform and an enhanced radial basis function neural network. Appl. Soft Comput. 118, 108542 (2022). https:\/\/doi.org\/10.1016\/j.asoc.2022.108542","journal-title":"Appl. Soft Comput."},{"issue":"12","key":"1473_CR7","doi-asserted-by":"publisher","first-page":"3450","DOI":"10.1109\/TBME.2012.2217493","volume":"59","author":"S Li","year":"2012","unstructured":"Li, S., Yin, H., Fang, L.: Group-sparse representation with dictionary learning for medical image denoising and fusion. IEEE Trans. Biomed. Eng. 59(12), 3450\u20133459 (2012). https:\/\/doi.org\/10.1109\/TBME.2012.2217493","journal-title":"IEEE Trans. Biomed. Eng."},{"key":"1473_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2022.103534","volume":"74","author":"W Li","year":"2022","unstructured":"Li, W., Li, R., Fu, J., Peng, X.: MSENet: a multi-scale enhanced network based on unique features guidance for medical image fusion. Biomed. Signal Process. Control 74, 103534 (2022). https:\/\/doi.org\/10.1016\/j.bspc.2022.103534","journal-title":"Biomed. Signal Process. Control"},{"issue":"11","key":"1473_CR9","doi-asserted-by":"publisher","first-page":"15017","DOI":"10.1007\/s11042-022-12189-4","volume":"81","author":"N Yu","year":"2022","unstructured":"Yu, N., Li, J., Hua, Z.: Decolorization algorithm based on contrast pyramid transform fusion. Multim. Tools Appl. 81(11), 15017\u201315039 (2022). https:\/\/doi.org\/10.1007\/s11042-022-12189-4","journal-title":"Multim. Tools Appl."},{"issue":"6","key":"1473_CR10","doi-asserted-by":"publisher","first-page":"1406","DOI":"10.37188\/OPE.20212906.1406","volume":"29","author":"J Lin","year":"2021","unstructured":"Lin, J., Liao, Y.: A novel image fusion method with fractional saliency detection and QFWA in NSST. Opt. Precis. Eng. 29(6), 1406\u20131419 (2021). https:\/\/doi.org\/10.37188\/OPE.20212906.1406","journal-title":"Opt. Precis. Eng."},{"issue":"5","key":"1473_CR11","doi-asserted-by":"publisher","first-page":"1308","DOI":"10.1007\/s10278-021-00554-y","volume":"35","author":"N Tawfik","year":"2022","unstructured":"Tawfik, N., Elnemr, H.A., Fakhr, M., Dessouky, M.I., El-Samie, F.E.A.: Multimodal medical image fusion using stacked auto-encoder in NSCT domain. J. Digit. Imaging 35(5), 1308\u20131325 (2022). https:\/\/doi.org\/10.1007\/s10278-021-00554-y","journal-title":"J. Digit. Imaging"},{"issue":"12","key":"1473_CR12","doi-asserted-by":"publisher","first-page":"3450","DOI":"10.1109\/tbme.2012.2217493","volume":"59","author":"S Li","year":"2012","unstructured":"Li, S., Yin, H., Fang, L.: Group-sparse representation with dictionary learning for medical image denoising and fusion. IEEE Trans. Biomed. Eng. 59(12), 3450\u20133459 (2012). https:\/\/doi.org\/10.1109\/tbme.2012.2217493","journal-title":"IEEE Trans. Biomed. Eng."},{"key":"1473_CR13","doi-asserted-by":"publisher","first-page":"195","DOI":"10.1016\/j.bspc.2017.02.005","volume":"34","author":"JJ Zong","year":"2017","unstructured":"Zong, J.J., Qiu, T.S.: Medical image fusion based on sparse representation of classified image patches. Biomed. Signal Process. Control 34, 195\u2013205 (2017). https:\/\/doi.org\/10.1016\/j.bspc.2017.02.005","journal-title":"Biomed. Signal Process. Control"},{"key":"1473_CR14","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1007\/s13131-015-0634-7","volume":"34","author":"M Liu","year":"2015","unstructured":"Liu, M., Dai, Y., Zhang, J., Zhang, X., Meng, J., Xie, Q.: PCA-based sea-ice image fusion of optical data by HIS transform and SAR data by wavelet transform. Acta Oceanol. Sin. 34, 59\u201367 (2015). https:\/\/doi.org\/10.1007\/s13131-015-0634-7","journal-title":"Acta Oceanol. Sin."},{"issue":"4","key":"1473_CR15","doi-asserted-by":"publisher","first-page":"746","DOI":"10.1016\/S0165-5728(98)00125-8","volume":"7","author":"S Rahmani","year":"2010","unstructured":"Rahmani, S., Strait, M., Merkurjev, D., Moeller, M., Wittman, T.: An adaptive IHS pan-sharpening method. IEEE Geosci. Remote Sens. Lett. 7(4), 746\u2013750 (2010). https:\/\/doi.org\/10.1016\/S0165-5728(98)00125-8","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"issue":"5","key":"1473_CR16","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.48550\/arXiv.1804.08361","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.J.: DenseFuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018). https:\/\/doi.org\/10.48550\/arXiv.1804.08361","journal-title":"IEEE Trans. Image Process."},{"key":"1473_CR17","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu, Y., Chen, X., Peng, H., Wang, Z.: Multi-focus image fusion with a deep convolutional neural network. Inform. Fusion 36, 191\u2013207 (2017). https:\/\/doi.org\/10.1016\/j.inffus.2016.12.001","journal-title":"Inform. Fusion"},{"key":"1473_CR18","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1016\/j.inffus.2021.06.001","volume":"76","author":"H Xu","year":"2021","unstructured":"Xu, H., Ma, J.: EMFusion: an unsupervised enhanced medical image fusion network. Inform. Fusion 76, 177\u2013186 (2021). https:\/\/doi.org\/10.1016\/j.inffus.2021.06.001","journal-title":"Inform. Fusion"},{"issue":"10","key":"1473_CR19","doi-asserted-by":"publisher","first-page":"2761","DOI":"10.1007\/s11263-021-01501-8","volume":"129","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Ma, J.: SDNet: A versatile squeeze-and-decomposition network for real-time image fusion. Int. J. Comput. Vision 129(10), 2761\u20132785 (2021). https:\/\/doi.org\/10.1007\/s11263-021-01501-8","journal-title":"Int. J. Comput. Vision"},{"key":"1473_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.meatsci.2022.108784","volume":"188","author":"HJ Lee","year":"2022","unstructured":"Lee, H.J., Koh, Y.J., Kim, Y.K., Lee, S.H., Lee, J.H., Seo, D.W.: MSENet: marbling score estimation network for automated assessment of Korean beef. Meat Sci. 188, 108784 (2022). https:\/\/doi.org\/10.1016\/j.meatsci.2022.108784","journal-title":"Meat Sci."},{"key":"1473_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2022.104402","volume":"80","author":"W Li","year":"2023","unstructured":"Li, W., Zhang, Y., Wang, G., Huang, Y., Li, R.: DFENet: a dual-branch feature enhanced network integrating transformers and convolutional feature learning for multimodal medical image fusion. Biomed. Signal Process. Control 80, 104402 (2023). https:\/\/doi.org\/10.1016\/j.bspc.2022.104402","journal-title":"Biomed. Signal Process. Control"},{"key":"1473_CR22","doi-asserted-by":"publisher","first-page":"16040","DOI":"10.1109\/ACCESS.2017.2735865","volume":"5","author":"A Dogra","year":"2017","unstructured":"Dogra, A., Goyal, B., Agrawal, S.: From multi-scale decomposition to non-multi-scale decomposition methods: a comprehensive survey of image fusion techniques and its applications. IEEE access 5, 16040\u201316067 (2017). https:\/\/doi.org\/10.1109\/ACCESS.2017.2735865","journal-title":"IEEE access"},{"issue":"7","key":"1473_CR23","doi-asserted-by":"publisher","first-page":"9277","DOI":"10.1007\/s11042-021-11549-w","volume":"81","author":"L Zhang","year":"2022","unstructured":"Zhang, L., Li, H., Zhu, R., Du, P.: An infrared and visible image fusion algorithm based on ResNet-152. Multim. Tools Appl. 81(7), 9277\u20139287 (2022). https:\/\/doi.org\/10.1007\/s11042-021-11549-w","journal-title":"Multim. Tools Appl."},{"issue":"6","key":"1473_CR24","doi-asserted-by":"publisher","first-page":"3360","DOI":"10.1109\/TCSVT.2021.3109895","volume":"32","author":"Z Wang","year":"2021","unstructured":"Wang, Z., Wang, J., Wu, Y., Xu, J., Zhang, X.: UNFusion: A unified multi-scale densely connected network for infrared and visible image fusion. IEEE Trans. Circuits Syst. Video Technol. 32(6), 3360\u20133374 (2021). https:\/\/doi.org\/10.1109\/TCSVT.2021.3109895","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"10","key":"1473_CR25","doi-asserted-by":"publisher","first-page":"4291","DOI":"10.48550\/arXiv.1902.02181","volume":"32","author":"A Galassi","year":"2020","unstructured":"Galassi, A., Lippi, M., Torroni, P.: Attention in natural language processing. IEEE Trans. Neural Netw. Learn. Syst. 32(10), 4291\u20134308 (2020). https:\/\/doi.org\/10.48550\/arXiv.1902.02181","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"1473_CR26","doi-asserted-by":"publisher","unstructured":"Cao, Y., Chen, K., Loy, C. C., Lin, D.: Prime sample attention in object detection. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 11583-11591).https:\/\/doi.org\/10.48550\/arXiv.1904.04821 (2020)","DOI":"10.48550\/arXiv.1904.04821"},{"key":"1473_CR27","doi-asserted-by":"publisher","first-page":"1140","DOI":"10.48550\/arXiv.2209.08575","volume":"35","author":"MH Guo","year":"2022","unstructured":"Guo, M.H., Lu, C.Z., Hou, Q., Liu, Z., Cheng, M.M., Hu, S.M.: Segnext: Rethinking convolutional attention design for semantic segmentation. Adv. Neural. Inf. Process. Syst. 35, 1140\u20131156 (2022). https:\/\/doi.org\/10.48550\/arXiv.2209.08575","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1473_CR28","doi-asserted-by":"publisher","first-page":"6816","DOI":"10.1109\/tip.2022.3216208","volume":"31","author":"Q Qi","year":"2022","unstructured":"Qi, Q., Li, K., Zheng, H., Gao, X., Hou, G., Sun, K.: SGUIE-Net: Semantic attention guided underwater image enhancement with multi-scale perception. IEEE Trans. Image Process. 31, 6816\u20136830 (2022). https:\/\/doi.org\/10.1109\/tip.2022.3216208","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"1473_CR29","doi-asserted-by":"publisher","first-page":"3207","DOI":"10.1007\/s11263-023-01843-5","volume":"131","author":"Y Mei","year":"2023","unstructured":"Mei, Y., Fan, Y., Zhang, Y., Yu, J., Zhou, Y., Liu, D., Fu, Y., Huang, T.S., Shi, H.: Pyramid attention network for image restoration. Int. J. Comput. Vision 131(12), 3207\u20133225 (2023). https:\/\/doi.org\/10.1007\/s11263-023-01843-5","journal-title":"Int. J. Comput. Vision"},{"key":"1473_CR30","doi-asserted-by":"publisher","first-page":"1949","DOI":"10.48550\/arXiv.2004.14582","volume":"30","author":"Z Zhang","year":"2021","unstructured":"Zhang, Z., Lin, Z., Xu, J., Jin, W.D., Lu, S.P., Fan, D.P.: Bilateral attention network for RGB-D salient object detection. IEEE Trans. Image Process. 30, 1949\u20131961 (2021). https:\/\/doi.org\/10.48550\/arXiv.2004.14582","journal-title":"IEEE Trans. Image Process."},{"issue":"8","key":"1473_CR31","doi-asserted-by":"publisher","first-page":"5026","DOI":"10.1109\/TCSVT.2022.3144455","volume":"32","author":"J Liu","year":"2022","unstructured":"Liu, J., Shang, J., Liu, R., Fan, X.: Attention-guided global-local adversarial learning for detail-preserving multi-exposure image fusion. IEEE Trans. Circuits Syst. Video Technol. 32(8), 5026\u20135040 (2022). https:\/\/doi.org\/10.1109\/TCSVT.2022.3144455","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1473_CR32","doi-asserted-by":"publisher","first-page":"33","DOI":"10.5815\/ijigsp.2021.04.03","volume":"13","author":"J Xu","year":"2021","unstructured":"Xu, J.: A review of self-supervised learning methods in the field of medical image analysis. Int. J. Image Graph. Signal Process. (IJIGSP) 13, 33\u201346 (2021). https:\/\/doi.org\/10.5815\/ijigsp.2021.04.03","journal-title":"Int. J. Image Graph. Signal Process. (IJIGSP)"},{"issue":"21","key":"1473_CR33","first-page":"58","volume":"13","author":"W Haobo","year":"2022","unstructured":"Haobo, W.: Comparative learning leads weak label learning new SOTA, and Zhejiang University\u2019s new research was selected as ICLR Oral. Heart 13(21), 58 (2022). (https:\/\/arxiv.org\/pdf\/2201.08984v2)","journal-title":"Heart"},{"issue":"1","key":"1473_CR34","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1186\/s40537-021-00434-w","volume":"8","author":"S Srivastava","year":"2021","unstructured":"Srivastava, S., Divekar, A.V., Anilkumar, C., Naik, I., Kulkarni, V., Pattabiraman, V.: Comparative analysis of deep learning image detection algorithms. J. Big Data 8(1), 66 (2021). https:\/\/doi.org\/10.1186\/s40537-021-00434-w","journal-title":"J. Big Data"},{"issue":"1","key":"1473_CR35","doi-asserted-by":"publisher","first-page":"179","DOI":"10.1007\/s00371-020-02010-4","volume":"38","author":"B Das","year":"2022","unstructured":"Das, B., Ebenezer, J.P., Mukhopadhyay, S.: A comparative study of single image fog removal methods. Vis. Comput. 38(1), 179\u2013195 (2022). https:\/\/doi.org\/10.1007\/s00371-020-02010-4","journal-title":"Vis. Comput."},{"issue":"12","key":"1473_CR36","doi-asserted-by":"publisher","first-page":"8476","DOI":"10.48550\/arXiv.2207.12767","volume":"32","author":"Y Shi","year":"2022","unstructured":"Shi, Y., Li, H., Zhang, S., Yang, Z., Wang, X.: Criteria comparative learning for real-scene image super-resolution. IEEE Trans. Circuits Syst. Video Technol. 32(12), 8476\u20138485 (2022). https:\/\/doi.org\/10.48550\/arXiv.2207.12767","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1473_CR37","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2211.10960","author":"J Liu","year":"2023","unstructured":"Liu, J., Lin, R., Wu, G., Liu, R., Luo, Z., Fan, X.: Coconet: Coupled contrastive learning network with multi-level feature ensemble for multi-modality image fusion. Int. J. Comput. Vis. (2023). https:\/\/doi.org\/10.48550\/arXiv.2211.10960","journal-title":"Int. J. Comput. Vis."},{"key":"1473_CR38","doi-asserted-by":"publisher","first-page":"389","DOI":"10.1016\/j.inffus.2022.12.002","volume":"92","author":"L Qu","year":"2023","unstructured":"Qu, L., Liu, S., Wang, M., Song, Z.: Rethinking multi-exposure image fusion with extreme and diverse exposure levels: a robust framework based on Fourier transform and contrastive learning. Inform. Fus. 92, 389\u2013403 (2023). https:\/\/doi.org\/10.1016\/j.inffus.2022.12.002","journal-title":"Inform. Fus."},{"key":"1473_CR39","doi-asserted-by":"publisher","unstructured":"Xu, H., Haochen, L., Ma, J.: Unsupervised multi-exposure image fusion breaking exposure limits via contrastive learning. In: Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 37, No. 3, pp. 3010-3017). https:\/\/doi.org\/10.1609\/aaai.v37i3.25404 (2023)","DOI":"10.1609\/aaai.v37i3.25404"},{"key":"1473_CR40","doi-asserted-by":"publisher","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 7132-7141). https:\/\/doi.org\/10.48550\/arXiv.1709.01507 (2018)","DOI":"10.48550\/arXiv.1709.01507"},{"issue":"10","key":"1473_CR41","doi-asserted-by":"publisher","first-page":"1736","DOI":"10.1093\/cid\/cix549","volume":"65","author":"DM Musher","year":"2017","unstructured":"Musher, D.M., Abers, M.S., Bartlett, J.G.: Evolving understanding of the causes of pneumonia in adults, with special attention to the role of pneumococcus. Clin. Infect. Dis. 65(10), 1736\u20131744 (2017). https:\/\/doi.org\/10.1093\/cid\/cix549","journal-title":"Clin. Infect. Dis."},{"issue":"1","key":"1473_CR42","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2020). https:\/\/doi.org\/10.1109\/TPAMI.2020.3012548","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1473_CR43","doi-asserted-by":"publisher","first-page":"5134","DOI":"10.1109\/TIP.2022.3193288","volume":"31","author":"W Tang","year":"2022","unstructured":"Tang, W., He, F., Liu, Y., Duan, Y.: MATR: Multimodal medical image fusion via multiscale adaptive transformer. IEEE Trans. Image Process. 31, 5134\u20135149 (2022). https:\/\/doi.org\/10.1109\/TIP.2022.3193288","journal-title":"IEEE Trans. Image Process."},{"key":"1473_CR44","doi-asserted-by":"publisher","DOI":"10.1016\/j.sigpro.2022.108637","volume":"200","author":"FG Veshki","year":"2022","unstructured":"Veshki, F.G., Ouzir, N., Vorobyov, S.A., Ollila, E.: Multimodal image fusion via coupled feature learning. Signal Process. 200, 108637 (2022). https:\/\/doi.org\/10.1016\/j.sigpro.2022.108637","journal-title":"Signal Process."},{"issue":"7","key":"1473_CR45","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma, J., Tang, L., Fan, F., Huang, J., Mei, X., Ma, Y.: SwinFusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA J. Autom. Sin. 9(7), 1200\u20131217 (2022). https:\/\/doi.org\/10.1109\/JAS.2022.105686","journal-title":"IEEE\/CAA J. Autom. Sin."},{"key":"1473_CR46","doi-asserted-by":"publisher","unstructured":"Huang, J., Li, X., Tan, H., Cheng, X.: Multimodal medical image fusion based on multichannel aggregated network. In International Conference on Image and Graphics (pp. 14-25). Cham: Springer Nature Switzerland. (2023) https:\/\/doi.org\/10.1016\/j.neucom.2022.01.059","DOI":"10.1016\/j.neucom.2022.01.059"},{"key":"1473_CR47","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2023.3293274","author":"J Li","year":"2023","unstructured":"Li, J., Liu, J., Zhou, S., Zhang, Q., Kasabov, N.K.: Gesenet: a general semantic-guided network with couple mask ensemble for medical image fusion. IEEE Trans. Neural Netw. Learn. Syst. (2023). https:\/\/doi.org\/10.1109\/TNNLS.2023.3293274","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"1473_CR48","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1806.07119","volume":"102","author":"H Li","year":"2019","unstructured":"Li, H., Wu, X.J., Durrani, T.S.: Infrared and visible image fusion with ResNet and zero-phase component analysis. Infrared Phys. Technol. 102, 103039 (2019). https:\/\/doi.org\/10.48550\/arXiv.1806.07119","journal-title":"Infrared Phys. Technol."},{"issue":"12","key":"1473_CR49","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.48550\/arXiv.2007.00328","volume":"69","author":"H Li","year":"2020","unstructured":"Li, H., Wu, X.J., Durrani, T.: NestFuse: An infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models. IEEE Trans. Instrum. Meas. 69(12), 9645\u20139656 (2020). https:\/\/doi.org\/10.48550\/arXiv.2007.00328","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"1473_CR50","doi-asserted-by":"publisher","DOI":"10.1016\/j.sigpro.2020.107734","volume":"177","author":"Z Zhao","year":"2020","unstructured":"Zhao, Z., Xu, S., Zhang, C., Liu, J., Zhang, J.: Bayesian fusion for infrared and visible images. Signal Process. 177, 107734 (2020). https:\/\/doi.org\/10.1016\/j.sigpro.2020.107734","journal-title":"Signal Process."},{"issue":"3","key":"1473_CR51","doi-asserted-by":"publisher","first-page":"1186","DOI":"10.48550\/arXiv.2005.05896","volume":"32","author":"Z Zhao","year":"2021","unstructured":"Zhao, Z., Xu, S., Zhang, J., Liang, C., Zhang, C., Liu, J.: Efficient and model-based infrared and visible image fusion via algorithm unrolling. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1186\u20131196 (2021). https:\/\/doi.org\/10.48550\/arXiv.2005.05896","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1473_CR52","doi-asserted-by":"publisher","first-page":"72","DOI":"10.48550\/arXiv.2103.04286","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Wu, X.J., Kittler, J.: RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inform. Fusion 73, 72\u201386 (2021). https:\/\/doi.org\/10.48550\/arXiv.2103.04286","journal-title":"Inform. Fusion"},{"key":"1473_CR53","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2022.104383","volume":"127","author":"W Xue","year":"2022","unstructured":"Xue, W., Wang, A., Zhao, L.: FLFuse-Net: A fast and lightweight infrared and visible image fusion network via feature flow and edge compensation for salient information. Infrared Phys. Technol. 127, 104383 (2022). https:\/\/doi.org\/10.1016\/j.infrared.2022.104383","journal-title":"Infrared Phys. Technol."},{"key":"1473_CR54","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","volume":"83","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Zhang, H., Jiang, X., Ma, J.: PIAFusion: A progressive infrared and visible image fusion network based on illumination aware. Inform. Fusion 83, 79\u201392 (2022). https:\/\/doi.org\/10.1016\/j.inffus.2022.03.007","journal-title":"Inform. Fusion"},{"key":"1473_CR55","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inform. Fusion 82, 28\u201342 (2022). https:\/\/doi.org\/10.1016\/j.inffus.2021.12.004","journal-title":"Inform. Fusion"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01473-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01473-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01473-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T18:11:47Z","timestamp":1730139107000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01473-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,9]]},"references-count":55,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["1473"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01473-y","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2024,9,9]]},"assertion":[{"value":"30 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 August 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 September 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"264"}}