{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,27]],"date-time":"2026-02-27T11:19:05Z","timestamp":1772191145751,"version":"3.50.1"},"reference-count":57,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2024,2,14]],"date-time":"2024-02-14T00:00:00Z","timestamp":1707868800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,2,14]],"date-time":"2024-02-14T00:00:00Z","timestamp":1707868800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62202416"],"award-info":[{"award-number":["62202416"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62162068"],"award-info":[{"award-number":["62162068"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,10]]},"DOI":"10.1007\/s00371-024-03273-x","type":"journal-article","created":{"date-parts":[[2024,2,14]],"date-time":"2024-02-14T13:02:31Z","timestamp":1707915751000},"page":"6739-6761","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["MVSFusion: infrared and visible image fusion method for multiple visual scenarios"],"prefix":"10.1007","volume":"40","author":[{"given":"Chengzhou","family":"Li","sequence":"first","affiliation":[]},{"given":"Kangjian","family":"He","sequence":"additional","affiliation":[]},{"given":"Dan","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Yueying","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Yiqiao","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,14]]},"reference":[{"key":"3273_CR1","doi-asserted-by":"publisher","first-page":"4353","DOI":"10.1007\/s00371-021-02300-5","volume":"38","author":"NS Jagtap","year":"2022","unstructured":"Jagtap, N.S., Thepade, S.D.: High-quality image multi-focus fusion to address ringing and blurring artifacts without loss of information. Vis. Comput. 38, 4353\u20134371 (2022). https:\/\/doi.org\/10.1007\/s00371-021-02300-5","journal-title":"Vis. Comput."},{"key":"3273_CR2","doi-asserted-by":"publisher","first-page":"2735","DOI":"10.1109\/TCYB.2019.2934823","volume":"51","author":"H Guo","year":"2021","unstructured":"Guo, H., Sheng, B., Li, P., Chen, C.L.P.: Multiview high dynamic range image synthesis using fuzzy broad learning system. IEEE Trans Cybern. 51, 2735\u20132747 (2021). https:\/\/doi.org\/10.1109\/TCYB.2019.2934823","journal-title":"IEEE Trans Cybern."},{"key":"3273_CR3","doi-asserted-by":"publisher","first-page":"4943","DOI":"10.1109\/TMM.2022.3185887","volume":"25","author":"K He","year":"2023","unstructured":"He, K., Zhang, X., Xu, D., Gong, J., Xie, L.: Fidelity-driven optimization reconstruction and details preserving guided fusion for multi-modality medical image. IEEE Trans. Multimed. 25, 4943\u20134957 (2023)","journal-title":"IEEE Trans. Multimed."},{"key":"3273_CR4","doi-asserted-by":"publisher","first-page":"640","DOI":"10.1109\/TCI.2020.2965304","volume":"6","author":"R Hou","year":"2020","unstructured":"Hou, R., Zhou, D., Nie, R., Liu, D., Xiong, L., Guo, Y., Yu, C.: VIF-Net: an unsupervised framework for infrared and visible image fusion. IEEE Trans. Comput. Imaging 6, 640\u2013651 (2020)","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"3273_CR5","doi-asserted-by":"publisher","unstructured":"Tan, A., Guo, T., Zhao, Y., Wang, Y., Li, X.: Object detection based on polarization image fusion and grouped convolutional attention network. Vis. Comput. 1\u201317 (2023).  https:\/\/doi.org\/10.1007\/s00371-023-03022-6","DOI":"10.1007\/s00371-023-03022-6"},{"key":"3273_CR6","doi-asserted-by":"publisher","first-page":"2725","DOI":"10.1007\/s00371-022-02488-0","volume":"39","author":"R Soroush","year":"2023","unstructured":"Soroush, R., Baleghi, Y.: NIR\/RGB image fusion for scene classification using deep neural networks. Vis. Comput. 39, 2725\u20132739 (2023). https:\/\/doi.org\/10.1007\/s00371-022-02488-0","journal-title":"Vis. Comput."},{"key":"3273_CR7","doi-asserted-by":"publisher","first-page":"103875","DOI":"10.1016\/j.dsp.2022.103875","volume":"133","author":"Z Ding","year":"2023","unstructured":"Ding, Z., Li, H., Zhou, D., Liu, Y., Hou, R.: Multi-spectral color vision fusion jointly with two-stream feature interaction and color transformation network. Digit. Signal Process. 133, 103875 (2023)","journal-title":"Digit. Signal Process."},{"key":"3273_CR8","doi-asserted-by":"publisher","unstructured":"Yu, C., Li, S., Feng, W., Zheng, T., Liu, S.: SACA-fusion: a low-light fusion architecture of infrared and visible images based on self-and cross-attention. Vis. Comput. 1\u201310 (2023). https:\/\/doi.org\/10.1007\/s00371-023-03037-z","DOI":"10.1007\/s00371-023-03037-z"},{"key":"3273_CR9","doi-asserted-by":"publisher","first-page":"4816","DOI":"10.1109\/TIP.2020.2976190","volume":"29","author":"J Li","year":"2020","unstructured":"Li, J., Guo, X., Lu, G., Zhang, B., Xu, Y., Wu, F., Zhang, D.: DRPL: deep regression pair learning for multi-focus image fusion. IEEE Trans. Image Process. 29, 4816\u20134831 (2020). https:\/\/doi.org\/10.1109\/TIP.2020.2976190","journal-title":"IEEE Trans. Image Process."},{"key":"3273_CR10","doi-asserted-by":"crossref","unstructured":"Lin, X., Li, J., Ma, Z., Li, H., Li, S., Xu, K., Lu, G., Zhang, D.: Learning modal-invariant and temporal-memory for video-based visible-infrared person re-identification. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 20973\u201320982 (2022)","DOI":"10.1109\/CVPR52688.2022.02030"},{"key":"3273_CR11","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft coco: Common objects in context. In: Computer Vision\u2013ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6\u201312, 2014, Proceedings, Part V 13. pp. 740\u2013755. Springer (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"3273_CR12","doi-asserted-by":"publisher","first-page":"2321","DOI":"10.1007\/s00371-022-02438-w","volume":"39","author":"R Lu","year":"2023","unstructured":"Lu, R., Gao, F., Yang, X., Fan, J., Li, D.: A novel infrared and visible image fusion method based on multi-level saliency integration. Vis. Comput. 39, 2321\u20132335 (2023)","journal-title":"Vis. Comput."},{"key":"3273_CR13","doi-asserted-by":"publisher","first-page":"4869","DOI":"10.1007\/s00371-022-02633-9","volume":"39","author":"J Liu","year":"2022","unstructured":"Liu, J., Jiang, Z., Wu, G., Liu, R., Fan, X.: A unified image fusion framework with flexible bilevel paradigm integration. Vis. Comput. 39, 4869\u20134886 (2022)","journal-title":"Vis. Comput."},{"key":"3273_CR14","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2022","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44, 502\u2013518 (2022). https:\/\/doi.org\/10.1109\/TPAMI.2020.3012548","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3273_CR15","doi-asserted-by":"publisher","first-page":"4801","DOI":"10.1007\/s00371-022-02628-6","volume":"39","author":"X Wang","year":"2022","unstructured":"Wang, X., Hua, Z., Li, J.: Cross-UNet: dual-branch infrared and visible image fusion framework based on cross-convolution and attention mechanism. Vis. Comput. 39, 4801\u20134818 (2022)","journal-title":"Vis. Comput."},{"key":"3273_CR16","doi-asserted-by":"publisher","first-page":"1181","DOI":"10.1007\/s00371-021-02396-9","volume":"39","author":"JA Aghamaleki","year":"2023","unstructured":"Aghamaleki, J.A., Ghorbani, A.: Image fusion using dual tree discrete wavelet transform and weights optimization. Vis. Comput. 39, 1181\u20131191 (2023)","journal-title":"Vis. Comput."},{"key":"3273_CR17","doi-asserted-by":"publisher","first-page":"107734","DOI":"10.1016\/j.sigpro.2020.107734","volume":"177","author":"Z Zhao","year":"2020","unstructured":"Zhao, Z., Xu, S., Zhang, C., Liu, J., Zhang, J.: Bayesian fusion for infrared and visible images. Signal Process. 177, 107734 (2020)","journal-title":"Signal Process."},{"key":"3273_CR18","doi-asserted-by":"publisher","first-page":"4249","DOI":"10.1007\/s00371-022-02588-x","volume":"39","author":"Q Xie","year":"2022","unstructured":"Xie, Q., Hu, J., Wang, X., Zhang, D., Qin, H.: Novel and fast EMD-based image fusion via morphological filter. Vis. Comput. 39, 4249\u20134265 (2022)","journal-title":"Vis. Comput."},{"key":"3273_CR19","doi-asserted-by":"publisher","first-page":"7033","DOI":"10.1364\/AO.52.007033","volume":"52","author":"L Li","year":"2013","unstructured":"Li, L., Li, H., Dang, E., Liu, B.: Compressive sensing method for recognizing cat-eye effect targets. Appl. Opt. 52, 7033\u20137039 (2013)","journal-title":"Appl. Opt."},{"key":"3273_CR20","doi-asserted-by":"publisher","first-page":"510","DOI":"10.1049\/el.2014.0180","volume":"50","author":"L Li","year":"2014","unstructured":"Li, L., Li, H., Li, T., Gao, F.: Infrared small target detection in compressive domain. Electron. Lett. 50, 510\u2013512 (2014)","journal-title":"Electron. Lett."},{"key":"3273_CR21","doi-asserted-by":"crossref","unstructured":"Hou, R., Ren, T., Wu, G.: MIRNet: a robust RGBT tracking jointly with multi-modal interaction and refinement. In: 2022 IEEE International Conference on Multimedia and Expo (ICME). pp. 1\u20136. IEEE (2022)","DOI":"10.1109\/ICME52920.2022.9860018"},{"key":"3273_CR22","doi-asserted-by":"publisher","first-page":"6723","DOI":"10.1007\/s00371-022-02759-w","volume":"39","author":"W Yin","year":"2023","unstructured":"Yin, W., He, K., Xu, D., Yue, Y., Luo, Y.: Adaptive low light visual enhancement and high-significant target detection for infrared and visible image fusion. Vis. Comput. 39, 6723\u20136742 (2023). https:\/\/doi.org\/10.1007\/s00371-022-02759-w","journal-title":"Vis. Comput."},{"key":"3273_CR23","doi-asserted-by":"publisher","first-page":"8114","DOI":"10.1007\/s10489-022-03952-z","volume":"53","author":"Z Ding","year":"2023","unstructured":"Ding, Z., Li, H., Zhou, D., Liu, Y., Hou, R.: A robust infrared and visible image fusion framework via multi-receptive-field attention and color visual perception. Appl. Intell. 53, 8114\u20138132 (2023)","journal-title":"Appl. Intell."},{"key":"3273_CR24","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen, J., Li, X., Luo, L., Mei, X., Ma, J.: Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 508, 64\u201378 (2020). https:\/\/doi.org\/10.1016\/j.ins.2019.08.066","journal-title":"Inf. Sci."},{"key":"3273_CR25","unstructured":"Li, H., Wu, X.-J.: Infrared and visible image fusion using latent low-rank representation. arXiv180408992. (2018)"},{"key":"3273_CR26","doi-asserted-by":"publisher","first-page":"4733","DOI":"10.1109\/TIP.2020.2975984","volume":"29","author":"H Li","year":"2020","unstructured":"Li, H., Wu, X.-J., Kittler, J.: MDLatLRR: A novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 29, 4733\u20134746 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"3273_CR27","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1016\/j.inffus.2021.02.008","volume":"71","author":"G Li","year":"2021","unstructured":"Li, G., Lin, Y., Qu, X.: An infrared and visible image fusion method based on multi-scale transformation and norm optimization. Inf. Fusion. 71, 109\u2013129 (2021). https:\/\/doi.org\/10.1016\/j.inffus.2021.02.008","journal-title":"Inf. Fusion."},{"key":"3273_CR28","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","volume":"82","author":"J Ma","year":"2017","unstructured":"Ma, J., Zhou, Z., Wang, B., Zong, H.: Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 82, 8\u201317 (2017). https:\/\/doi.org\/10.1016\/j.infrared.2017.02.005","journal-title":"Infrared Phys. Technol."},{"key":"3273_CR29","doi-asserted-by":"publisher","first-page":"477","DOI":"10.1016\/j.inffus.2022.10.034","volume":"91","author":"L Tang","year":"2023","unstructured":"Tang, L., Xiang, X., Zhang, H., Gong, M., Ma, J.: DIVFusion: darkness-free infrared and visible image fusion. Inf. Fusion. 91, 477\u2013493 (2023)","journal-title":"Inf. Fusion."},{"key":"3273_CR30","doi-asserted-by":"publisher","first-page":"2761","DOI":"10.1007\/s11263-021-01501-8","volume":"129","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Ma, J.: SDNet: a versatile squeeze-and-decomposition network for real-time image fusion. Int. J. Comput. Vis. 129, 2761\u20132785 (2021)","journal-title":"Int. J. Comput. Vis."},{"key":"3273_CR31","doi-asserted-by":"crossref","unstructured":"Wang, D., Liu, J., Fan, X., Liu, R.: Unsupervised misaligned infrared and visible image fusion via cross-modality image generation and registration. arXiv220511876. (2022)","DOI":"10.24963\/ijcai.2022\/487"},{"key":"3273_CR32","doi-asserted-by":"publisher","first-page":"824","DOI":"10.1109\/TCI.2021.3100986","volume":"7","author":"H Xu","year":"2021","unstructured":"Xu, H., Zhang, H., Ma, J.: Classification saliency-based rule for visible and infrared image fusion. IEEE Trans. Comput. Imaging. 7, 824\u2013836 (2021)","journal-title":"IEEE Trans. Comput. Imaging."},{"key":"3273_CR33","doi-asserted-by":"publisher","first-page":"174","DOI":"10.1016\/j.inffus.2022.12.022","volume":"93","author":"Z Zhou","year":"2023","unstructured":"Zhou, Z., Fei, E., Miao, L., Yang, R.: A perceptual framework for infrared\u2013visible image fusion based on multiscale structure decomposition and biological vision. Inf. Fusion. 93, 174\u2013191 (2023)","journal-title":"Inf. Fusion."},{"key":"3273_CR34","doi-asserted-by":"publisher","first-page":"3064","DOI":"10.1364\/AO.58.003064","volume":"58","author":"W Tan","year":"2019","unstructured":"Tan, W., Zhou, H., Song, J., Li, H., Yu, Y., Du, J.: Infrared and visible image perceptive fusion through multi-level Gaussian curvature filtering image decomposition. Appl. Opt. 58, 3064 (2019). https:\/\/doi.org\/10.1364\/AO.58.003064","journal-title":"Appl. Opt."},{"key":"3273_CR35","doi-asserted-by":"publisher","first-page":"118631","DOI":"10.1016\/j.eswa.2022.118631","volume":"211","author":"C Guo","year":"2023","unstructured":"Guo, C., Fan, D., Jiang, Z., Zhang, D.: MDFN: mask deep fusion network for visible and infrared image fusion without reference ground-truth. Expert Syst. Appl. 211, 118631 (2023)","journal-title":"Expert Syst. Appl."},{"key":"3273_CR36","doi-asserted-by":"publisher","first-page":"22511","DOI":"10.1007\/s00521-023-08916-z","volume":"35","author":"C Li","year":"2023","unstructured":"Li, C., He, K., Xu, D., Tao, D., Lin, X., Shi, H., Yin, W.: Superpixel-based adaptive salient region analysis for infrared and visible image fusion. Neural Comput. Appl. 35, 22511\u201322529 (2023)","journal-title":"Neural Comput. Appl."},{"key":"3273_CR37","doi-asserted-by":"publisher","first-page":"104041","DOI":"10.1016\/j.infrared.2022.104041","volume":"121","author":"W Yin","year":"2022","unstructured":"Yin, W., He, K., Xu, D., Luo, Y., Gong, J.: Significant target analysis and detail preserving based infrared and visible image fusion. Infrared Phys. Technol. 121, 104041 (2022)","journal-title":"Infrared Phys. Technol."},{"key":"3273_CR38","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: FusionGAN: a generative adversarial network for infrared and visible image fusion. Inf. Fusion. 48, 11\u201326 (2019)","journal-title":"Inf. Fusion."},{"key":"3273_CR39","doi-asserted-by":"publisher","first-page":"26","DOI":"10.1016\/j.inffus.2023.02.011","volume":"95","author":"H Li","year":"2023","unstructured":"Li, H., Zhao, J., Li, J., Yu, Z., Lu, G.: Feature dynamic alignment and refinement for infrared-visible image fusion: translation robust fusion. Inf Fusion. 95, 26\u201341 (2023). https:\/\/doi.org\/10.1016\/j.inffus.2023.02.011","journal-title":"Inf Fusion."},{"key":"3273_CR40","doi-asserted-by":"publisher","first-page":"268","DOI":"10.1016\/J.INFFUS.2022.12.005","volume":"92","author":"M Han","year":"2023","unstructured":"Han, M., Yu, K., Qiu, J., Li, H., Wu, D., Rao, Y., Yang, Y., Xing, L., Bai, H., Zhou, C.: Boosting target-level infrared and visible image fusion with regional information coordination. Inf Fusion. 92, 268\u2013288 (2023). https:\/\/doi.org\/10.1016\/J.INFFUS.2022.12.005","journal-title":"Inf Fusion."},{"key":"3273_CR41","doi-asserted-by":"publisher","first-page":"769","DOI":"10.1109\/TCI.2023.3304471","volume":"9","author":"X Wang","year":"2023","unstructured":"Wang, X., Guan, Z., Qian, W., Cao, J., Wang, C., Yang, C.: Contrast saliency information guided infrared and visible image fusion. IEEE Trans. Comput. Imaging. 9, 769\u2013780 (2023).  https:\/\/doi.org\/10.1109\/TCI.2023.3304471","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"3273_CR42","doi-asserted-by":"crossref","unstructured":"Liu, J., Fan, X., Huang, Z., Wu, G., Liu, R., Zhong, W., Luo, Z.: Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 5802\u20135811 (2022)","DOI":"10.1109\/CVPR52688.2022.00571"},{"key":"3273_CR43","doi-asserted-by":"crossref","unstructured":"Takumi, K., Watanabe, K., Ha, Q., Tejero-De-Pablos, A., Ushiku, Y., Harada, T.: Multispectral object detection for autonomous vehicles. In: Proceedings of the on Thematic Workshops of ACM Multimedia 2017. pp. 35\u201343 (2017)","DOI":"10.1145\/3126686.3126727"},{"key":"3273_CR44","doi-asserted-by":"crossref","unstructured":"Lee, H., Jeon, J., Kim, J., Lee, S.: Structure-texture decomposition of images with interval gradient. In: Computer Graphics Forum. 36(6), 262\u2013274 (2017)","DOI":"10.1111\/cgf.12875"},{"key":"3273_CR45","doi-asserted-by":"publisher","first-page":"308","DOI":"10.1049\/el:20000267","volume":"36","author":"CS Xydeas","year":"2000","unstructured":"Xydeas, C.S., Petrovic, V.: Others: objective image fusion performance measure. Electron. Lett. 36, 308\u2013309 (2000)","journal-title":"Electron. Lett."},{"key":"3273_CR46","doi-asserted-by":"publisher","first-page":"430","DOI":"10.1109\/TIP.2005.859378","volume":"15","author":"HR Sheikh","year":"2006","unstructured":"Sheikh, H.R., Bovik, A.C.: Image information and visual quality. IEEE Trans. Image Process. 15, 430\u2013444 (2006)","journal-title":"IEEE Trans. Image Process."},{"key":"3273_CR47","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1049\/el:20020212","volume":"38","author":"G Qu","year":"2002","unstructured":"Qu, G., Zhang, D., Yan, P.: Information measure for performance of image fusion. Electron. Lett. 38, 1 (2002)","journal-title":"Electron. Lett."},{"key":"3273_CR48","doi-asserted-by":"publisher","first-page":"2117","DOI":"10.1109\/TIP.2005.859389","volume":"14","author":"HR Sheikh","year":"2005","unstructured":"Sheikh, H.R., Bovik, A.C., De Veciana, G.: An information fidelity criterion for image quality assessment using natural scene statistics. IEEE Trans. Image Process. 14, 2117\u20132128 (2005)","journal-title":"IEEE Trans. Image Process."},{"key":"3273_CR49","doi-asserted-by":"crossref","unstructured":"Ha, Q., Watanabe, K., Karasawa, T., Ushiku, Y., Harada, T.: MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems, IROS 2017, Vancouver, BC, Canada, September 24\u201328, 2017. pp. 5108\u20135115. IEEE (2017)","DOI":"10.1109\/IROS.2017.8206396"},{"key":"3273_CR50","doi-asserted-by":"publisher","first-page":"2274","DOI":"10.1109\/TPAMI.2012.120","volume":"34","author":"R Achanta","year":"2012","unstructured":"Achanta, R., Shaji, A., Smith, K., Lucchi, A., Fua, P., S\u00fcsstrunk, S.: SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 34, 2274\u20132282 (2012)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3273_CR51","doi-asserted-by":"crossref","unstructured":"Xu, H., Ma, J., Le, Z., Jiang, J., Guo, X.: FusionDN: a unified densely connected network for image fusion. In: The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7\u201312, 2020. pp. 12484\u201312491. AAAI Press (2020)","DOI":"10.1609\/aaai.v34i07.6936"},{"key":"3273_CR52","unstructured":"Toet, A.: TNO image fusion dataset. figshare. Dataset (2014)"},{"key":"3273_CR53","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.-J.: DenseFuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28, 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"3273_CR54","doi-asserted-by":"crossref","unstructured":"Lin, X., Sun, S., Huang, W., Sheng, B., Li, P., Feng, D.D.: EAPT: efficient attention pyramid transformer for image processing. IEEE Trans. Multimed. 25, 50\u201361 (2023)","DOI":"10.1109\/TMM.2021.3120873"},{"key":"3273_CR55","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"3273_CR56","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention\u2013MICCAI 2015: 18th International Conference, Munich, Germany, October 5\u20139, 2015, Proceedings, Part III 18. pp. 234\u2013241. Springer (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"3273_CR57","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03273-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-024-03273-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03273-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T10:04:17Z","timestamp":1727690657000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-024-03273-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,2,14]]},"references-count":57,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["3273"],"URL":"https:\/\/doi.org\/10.1007\/s00371-024-03273-x","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,2,14]]},"assertion":[{"value":"10 January 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 February 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there is no conflict of interest regarding the publication of the article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}