{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T02:55:09Z","timestamp":1775530509678,"version":"3.50.1"},"reference-count":56,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2023,1,16]],"date-time":"2023-01-16T00:00:00Z","timestamp":1673827200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,16]],"date-time":"2023-01-16T00:00:00Z","timestamp":1673827200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62202416"],"award-info":[{"award-number":["62202416"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62162068"],"award-info":[{"award-number":["62162068"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2023,12]]},"DOI":"10.1007\/s00371-022-02759-w","type":"journal-article","created":{"date-parts":[[2023,1,16]],"date-time":"2023-01-16T18:02:52Z","timestamp":1673892172000},"page":"6723-6742","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":30,"title":["Adaptive low light visual enhancement and high-significant target detection for infrared and visible image fusion"],"prefix":"10.1007","volume":"39","author":[{"given":"Wenxia","family":"Yin","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6207-9728","authenticated-orcid":false,"given":"Kangjian","family":"He","sequence":"additional","affiliation":[]},{"given":"Dan","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Yingying","family":"Yue","sequence":"additional","affiliation":[]},{"given":"Yueying","family":"Luo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,1,16]]},"reference":[{"key":"2759_CR1","first-page":"1","volume":"1","author":"R Lu","year":"2022","unstructured":"Lu, R., Gao, F., Yang, X., Fan, J., Li, D.: A novel infrared and visible image fusion method based on multi-level saliency integration. Vis. Comput. 1, 1\u201315 (2022)","journal-title":"Vis. Comput."},{"key":"2759_CR2","first-page":"1","volume":"1","author":"X Wang","year":"2022","unstructured":"Wang, X., Hua, Z., Li, J.: Cross-UNet: dual-branch infrared and visible image fusion framework based on cross-convolution and attention mechanism. Vis. Comput. 1, 1\u201318 (2022)","journal-title":"Vis. Comput."},{"key":"2759_CR3","first-page":"1","volume":"1","author":"J Liu","year":"2022","unstructured":"Liu, J., Jiang, Z., Wu, G., Liu, R., Fan, X.: A unified image fusion framework with flexible bilevel paradigm integration. Vis. Comput. 1, 1\u201318 (2022)","journal-title":"Vis. Comput."},{"key":"2759_CR4","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"J Ma","year":"2019","unstructured":"Ma, J., Ma, Y., Li, C.: Infrared and visible image fusion methods and applications: a survey. Inf. Fusion. 45, 153\u2013178 (2019). https:\/\/doi.org\/10.1016\/j.inffus.2018.02.004","journal-title":"Inf. Fusion."},{"key":"2759_CR5","doi-asserted-by":"crossref","unstructured":"Zhang, X., Ye, P., Xiao, G.: VIFB: A visible and infrared image fusion benchmark. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR Workshops 2020, Seattle, WA, USA, June 14\u201319, 2020. pp. 468\u2013478. Computer Vision Foundation\/IEEE (2020)","DOI":"10.1109\/CVPRW50498.2020.00060"},{"key":"2759_CR6","first-page":"1","volume":"1","author":"NS Jagtap","year":"2021","unstructured":"Jagtap, N.S., Thepade, S.D.: High-quality image multi-focus fusion to address ringing and blurring artifacts without loss of information. Vis. Comput. 1, 1\u201319 (2021)","journal-title":"Vis. Comput."},{"key":"2759_CR7","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-021-02396-9","author":"JA Aghamaleki","year":"2022","unstructured":"Aghamaleki, J.A., Ghorbani, A.: Image fusion using dual tree discrete wavelet transform and weights optimization. Vis. Comput. (2022). https:\/\/doi.org\/10.1007\/s00371-021-02396-9","journal-title":"Vis. Comput."},{"key":"2759_CR8","doi-asserted-by":"publisher","first-page":"1233","DOI":"10.1007\/s00371-021-02079-5","volume":"37","author":"C Wang","year":"2021","unstructured":"Wang, C., He, C., Xu, M.: Fast exposure fusion of detail enhancement for brightest and darkest regions. Vis. Comput. 37, 1233\u20131243 (2021)","journal-title":"Vis. Comput."},{"key":"2759_CR9","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2021.3066467","volume":"70","author":"K He","year":"2021","unstructured":"He, K., Gong, J., Xie, L., Zhang, X., Xu, D.: Regions preserving edge enhancement for multisensor-based medical image fusion. IEEE Trans. Instrum. Meas. 70, 1\u201313 (2021). https:\/\/doi.org\/10.1109\/TIM.2021.3066467","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"2759_CR10","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2021.3069116","volume":"60","author":"Y Chen","year":"2022","unstructured":"Chen, Y., Shi, K., Ge, Y., Zhou, Y.: Spatiotemporal remote sensing image fusion using multiscale two-stream convolutional neural networks. IEEE Trans. Geosci. Remote. Sens. 60, 1\u201312 (2022). https:\/\/doi.org\/10.1109\/TGRS.2021.3069116","journal-title":"IEEE Trans. Geosci. Remote. Sens."},{"key":"2759_CR11","doi-asserted-by":"publisher","first-page":"168914","DOI":"10.1016\/j.ijleo.2022.168914","volume":"1","author":"Y Luo","year":"2022","unstructured":"Luo, Y., He, K., Xu, D., Yin, W., Liu, W.: Infrared and visible image fusion based on visibility enhancement and hybrid multiscale decomposition. Optik 1, 168914 (2022)","journal-title":"Optik"},{"key":"2759_CR12","doi-asserted-by":"publisher","DOI":"10.1117\/1.JEI.31.1.013032","volume":"31","author":"Y Luo","year":"2022","unstructured":"Luo, Y., He, K., Xu, D., Yin, W.: Infrared and visible image fusion based on visibility enhancement and norm optimization low-rank representation. J. Electron. Imaging 31, 013032 (2022)","journal-title":"J. Electron. Imaging"},{"key":"2759_CR13","first-page":"1","volume":"1","author":"W Yin","year":"2022","unstructured":"Yin, W., He, K., Xu, D., Luo, Y., Gong, J.: Adaptive enhanced infrared and visible image fusion using hybrid decomposition and coupled dictionary. Neural Comput. Appl. 1, 1\u201319 (2022)","journal-title":"Neural Comput. Appl."},{"key":"2759_CR14","first-page":"1","volume":"1","author":"R Soroush","year":"2022","unstructured":"Soroush, R., Baleghi, Y.: NIR\/RGB image fusion for scene classification using deep neural networks. Vis. Comput. 1, 1\u201315 (2022)","journal-title":"Vis. Comput."},{"key":"2759_CR15","doi-asserted-by":"publisher","first-page":"259","DOI":"10.1016\/S1566-2535(03)00046-0","volume":"4","author":"G Piella","year":"2003","unstructured":"Piella, G.: A general framework for multiresolution image fusion: from pixels to regions. Inf. Fusion. 4, 259\u2013280 (2003). https:\/\/doi.org\/10.1016\/S1566-2535(03)00046-0","journal-title":"Inf. Fusion."},{"key":"2759_CR16","doi-asserted-by":"publisher","first-page":"478","DOI":"10.1016\/j.infrared.2017.07.010","volume":"85","author":"X Jin","year":"2017","unstructured":"Jin, X., Jiang, Q., Yao, S., Zhou, D., Nie, R., Hai, J., He, K.: A survey of infrared and visual image fusion methods. Infrar. Phys. Technol. 85, 478\u2013501 (2017)","journal-title":"Infrar. Phys. Technol."},{"key":"2759_CR17","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen, J., Li, X., Luo, L., Mei, X., Ma, J.: Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 508, 64\u201378 (2020). https:\/\/doi.org\/10.1016\/j.ins.2019.08.066","journal-title":"Inf. Sci."},{"key":"2759_CR18","first-page":"57","volume":"28","author":"L Zhan","year":"2017","unstructured":"Zhan, L., Zhuang, Y., Huang, L.: Infrared and visible images fusion method based on discrete wavelet transform. J. Comput. 28, 57\u201371 (2017)","journal-title":"J. Comput."},{"key":"2759_CR19","doi-asserted-by":"publisher","first-page":"5576","DOI":"10.1007\/s00034-019-01131-z","volume":"38","author":"DP Bavirisetti","year":"2019","unstructured":"Bavirisetti, D.P., Xiao, G., Zhao, J., Dhuli, R., Liu, G.: Multi-scale guided image and video fusion: a fast and efficient approach. Circuits Syst. Signal Process. 38, 5576\u20135605 (2019). https:\/\/doi.org\/10.1007\/s00034-019-01131-z","journal-title":"Circuits Syst. Signal Process."},{"key":"2759_CR20","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1109\/JSEN.2015.2478655","volume":"16","author":"DP Bavirisetti","year":"2015","unstructured":"Bavirisetti, D.P., Dhuli, R.: Fusion of infrared and visible sensor images based on anisotropic diffusion and Karhunen-Loeve transform. IEEE Sens. J. 16, 203\u2013209 (2015)","journal-title":"IEEE Sens. J."},{"key":"2759_CR21","doi-asserted-by":"crossref","unstructured":"Bavirisetti, D.P., Xiao, G., Liu, G.: Multi-sensor image fusion based on fourth order partial differential equations. In: 20th International Conference on Information Fusion, FUSION 2017, Xi\u2019an, China, July 10\u201313, 2017, pp. 1\u20139. IEEE (2017)","DOI":"10.23919\/ICIF.2017.8009719"},{"key":"2759_CR22","doi-asserted-by":"publisher","first-page":"1235","DOI":"10.1007\/s11042-018-6229-5","volume":"78","author":"R Bashir","year":"2019","unstructured":"Bashir, R., Junejo, R., Qadri, N.N., Fleury, M., Qadri, M.Y.: SWT and PCA image fusion methods for multi-modal imagery. Multim. Tools Appl. 78, 1235\u20131263 (2019). https:\/\/doi.org\/10.1007\/s11042-018-6229-5","journal-title":"Multim. Tools Appl."},{"key":"2759_CR23","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1016\/j.inffus.2005.09.001","volume":"8","author":"N Mitianoudis","year":"2007","unstructured":"Mitianoudis, N., Stathaki, T.: Pixel-based and region-based image fusion schemes using ICA bases. Inf. Fusion. 8, 131\u2013142 (2007). https:\/\/doi.org\/10.1016\/j.inffus.2005.09.001","journal-title":"Inf. Fusion."},{"key":"2759_CR24","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1016\/j.inffus.2021.02.008","volume":"71","author":"G Li","year":"2021","unstructured":"Li, G., Lin, Y., Qu, X.: An infrared and visible image fusion method based on multi-scale transformation and norm optimization. Inf. Fusion. 71, 109\u2013129 (2021)","journal-title":"Inf. Fusion."},{"key":"2759_CR25","doi-asserted-by":"crossref","unstructured":"Yin, W., He, K., Xu, D., Luo, Y., Gong, J.: Significant target analysis and detail preserving based infrared and visible image fusion. Infrar. Phys. Technol. 104041 (2022)","DOI":"10.1016\/j.infrared.2022.104041"},{"key":"2759_CR26","doi-asserted-by":"publisher","first-page":"1193","DOI":"10.1007\/s11760-013-0556-9","volume":"9","author":"BS Kumar","year":"2015","unstructured":"Kumar, B.S.: Image fusion based on pixel significance using cross bilateral filter. SIViP 9, 1193\u20131204 (2015)","journal-title":"SIViP"},{"key":"2759_CR27","doi-asserted-by":"publisher","first-page":"2864","DOI":"10.1109\/TIP.2013.2244222","volume":"22","author":"S Li","year":"2013","unstructured":"Li, S., Kang, X., Hu, J.: Image fusion with guided filtering. IEEE Trans. Image Process. 22, 2864\u20132875 (2013)","journal-title":"IEEE Trans. Image Process."},{"key":"2759_CR28","doi-asserted-by":"publisher","first-page":"107734","DOI":"10.1016\/j.sigpro.2020.107734","volume":"177","author":"Z Zhao","year":"2020","unstructured":"Zhao, Z., Xu, S., Zhang, C., Liu, J., Zhang, J.: Bayesian fusion for infrared and visible images. Signal Process. 177, 107734 (2020). https:\/\/doi.org\/10.1016\/j.sigpro.2020.107734","journal-title":"Signal Process."},{"key":"2759_CR29","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1016\/j.neucom.2022.09.157","volume":"514","author":"C Panigrahy","year":"2022","unstructured":"Panigrahy, C., Seal, A., Mahato, N.K.: Parameter adaptive unit-linking dual-channel PCNN based infrared and visible image fusion. Neurocomputing 514, 21\u201338 (2022)","journal-title":"Neurocomputing"},{"key":"2759_CR30","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Wu, X.-J., Kittler, J.: RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf. Fusion. 73, 72\u201386 (2021). https:\/\/doi.org\/10.1016\/j.inffus.2021.02.023","journal-title":"Inf. Fusion."},{"key":"2759_CR31","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: FusionGAN: a generative adversarial network for infrared and visible image fusion. Inf. Fusion. 48, 11\u201326 (2019). https:\/\/doi.org\/10.1016\/j.inffus.2018.09.004","journal-title":"Inf. Fusion."},{"key":"2759_CR32","doi-asserted-by":"publisher","first-page":"7458","DOI":"10.1109\/JSEN.2019.2921803","volume":"21","author":"Q Li","year":"2019","unstructured":"Li, Q., Lu, L., Li, Z., Wu, W., Liu, Z., Jeon, G., Yang, X.: Coupled GAN with relativistic discriminators for infrared and visible images fusion. IEEE Sens. J. 21, 7458\u20137467 (2019)","journal-title":"IEEE Sens. J."},{"key":"2759_CR33","first-page":"1","volume":"60","author":"Z Tan","year":"2021","unstructured":"Tan, Z., Gao, M., Li, X., Jiang, L.: A flexible reference-insensitive spatiotemporal fusion model for remote sensing images using conditional generative adversarial network. IEEE Trans. Geosci. Remote Sens. 60, 1\u201313 (2021)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"2759_CR34","doi-asserted-by":"publisher","first-page":"110","DOI":"10.1016\/j.inffus.2021.02.019","volume":"72","author":"Y Fu","year":"2021","unstructured":"Fu, Y., Wu, X.-J., Durrani, T.S.: Image fusion based on generative adversarial network consistent with perception. Inf. Fusion. 72, 110\u2013125 (2021). https:\/\/doi.org\/10.1016\/j.inffus.2021.02.019","journal-title":"Inf. Fusion."},{"key":"2759_CR35","doi-asserted-by":"publisher","first-page":"575","DOI":"10.1016\/j.asoc.2016.10.012","volume":"52","author":"L Maurya","year":"2017","unstructured":"Maurya, L., Mahapatra, P.K., Kumar, A.: A social spider optimized image fusion approach for contrast enhancement and brightness preservation. Appl. Soft Comput. 52, 575\u2013592 (2017). https:\/\/doi.org\/10.1016\/j.asoc.2016.10.012","journal-title":"Appl. Soft Comput."},{"key":"2759_CR36","doi-asserted-by":"publisher","first-page":"5022","DOI":"10.1109\/TIP.2020.2974060","volume":"29","author":"J Xu","year":"2020","unstructured":"Xu, J., Hou, Y., Ren, D., Liu, L., Zhu, F., Yu, M., Wang, H., Shao, L.: STAR: A Structure and Texture Aware Retinex Model. IEEE Trans. Image Process. 29, 5022\u20135037 (2020). https:\/\/doi.org\/10.1109\/TIP.2020.2974060","journal-title":"IEEE Trans. Image Process."},{"key":"2759_CR37","unstructured":"Tomasi, C., Manduchi, R.: Bilateral Filtering for Gray and Color Images. In: Proceedings of the Sixth International Conference on Computer Vision (ICCV-98), Bombay, India, January 4\u20137, 1998. pp. 839\u2013846. IEEE Computer Society (1998)"},{"key":"2759_CR38","doi-asserted-by":"crossref","unstructured":"Ghosh, S., Chaudhury, K.N.: Fast Bright-Pass Bilateral Filtering for Low-Light Enhancement. In: 2019 IEEE International Conference on Image Processing, ICIP 2019, Taipei, Taiwan, September 22\u201325, 2019. pp. 205\u2013209. IEEE (2019)","DOI":"10.1109\/ICIP.2019.8802986"},{"key":"2759_CR39","first-page":"1","volume":"60","author":"X Kong","year":"2021","unstructured":"Kong, X., Yang, C., Cao, S., Li, C., Peng, Z.: Infrared small target detection via nonconvex tensor fibered rank approximation. IEEE Trans. Geosci. Remote Sens. 60, 1\u201321 (2021)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"2759_CR40","doi-asserted-by":"publisher","DOI":"10.1016\/j.sigpro.2020.107645","volume":"174","author":"Y Chen","year":"2020","unstructured":"Chen, Y., Li, J., Zhou, Y.: Hyperspectral image denoising by total variation-regularized bilinear factorization. Signal Process. 174, 107645 (2020)","journal-title":"Signal Process."},{"key":"2759_CR41","doi-asserted-by":"publisher","first-page":"734","DOI":"10.1109\/TGRS.2019.2940534","volume":"58","author":"Y-B Zheng","year":"2019","unstructured":"Zheng, Y.-B., Huang, T.-Z., Zhao, X.-L., Jiang, T.-X., Ma, T.-H., Ji, T.-Y.: Mixed noise removal in hyperspectral image via low-fibered-rank regularization. IEEE Trans. Geosci. Remote Sens. 58, 734\u2013749 (2019)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"2759_CR42","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1007\/BF01581204","volume":"55","author":"J Eckstein","year":"1992","unstructured":"Eckstein, J., Bertsekas, D.P.: On the Douglas\u2014Rachford splitting method and the proximal point algorithm for maximal monotone operators. Math. Program. 55, 293\u2013318 (1992)","journal-title":"Math. Program."},{"key":"2759_CR43","doi-asserted-by":"publisher","first-page":"145","DOI":"10.1016\/S1364-6613(03)00055-X","volume":"7","author":"S Dehaene","year":"2003","unstructured":"Dehaene, S.: The neural basis of the Weber-Fechner law: a logarithmic mental number line. Trends Cogn. Sci. 7, 145\u2013147 (2003)","journal-title":"Trends Cogn. Sci."},{"key":"2759_CR44","unstructured":"Ying, Z., Li, G., Ren, Y., Wang, R., Wang, W.: A New Image Contrast Enhancement Algorithm Using Exposure Fusion Framework. In: Felsberg, M., Heyden, A., and Kr\u00fcger, N. (eds.) Computer Analysis of Images and Patterns - 17th International Conference, CAIP 2017, Ystad, Sweden, August 22\u201324, 2017, Proceedings, Part II. pp. 36\u201346. Springer (2017)"},{"key":"2759_CR45","doi-asserted-by":"publisher","first-page":"430","DOI":"10.1109\/TIP.2005.859378","volume":"15","author":"HR Sheikh","year":"2006","unstructured":"Sheikh, H.R., Bovik, A.C.: Image information and visual quality. IEEE Trans. Image Process. 15, 430\u2013444 (2006). https:\/\/doi.org\/10.1109\/TIP.2005.859378","journal-title":"IEEE Trans. Image Process."},{"key":"2759_CR46","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu, Y., Chen, X., Peng, H., Wang, Z.: Multi-focus image fusion with a deep convolutional neural network. Inf. Fus. 36, 191\u2013207 (2017)","journal-title":"Inf. Fus."},{"key":"2759_CR47","doi-asserted-by":"publisher","first-page":"1397","DOI":"10.1109\/TPAMI.2012.213","volume":"35","author":"K He","year":"2012","unstructured":"He, K., Sun, J., Tang, X.: Guided image filtering. IEEE Trans. Pattern Anal. Mach. Intell. 35, 1397\u20131409 (2012)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2759_CR48","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13, 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"2759_CR49","doi-asserted-by":"crossref","unstructured":"Qin, X., Zhang, Z.V., Huang, C., Gao, C., Dehghan, M., J\u00e4gersand, M.: BASNet: Boundary-Aware Salient Object Detection. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16\u201320, 2019. pp. 7479\u20137489. Computer Vision Foundation\/IEEE (2019)","DOI":"10.1109\/CVPR.2019.00766"},{"key":"2759_CR50","doi-asserted-by":"crossref","unstructured":"Liang, Z., Xu, J., Zhang, D., Cao, Z., Zhang, L.: A Hybrid l1-l0 Layer Decomposition Model for Tone Mapping. In: 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18\u201322, 2018. pp. 4758\u20134766. Computer Vision Foundation \/ IEEE Computer Society (2018)","DOI":"10.1109\/CVPR.2018.00500"},{"key":"2759_CR51","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1016\/j.inffus.2014.09.004","volume":"24","author":"Y Liu","year":"2015","unstructured":"Liu, Y., Liu, S., Wang, Z.: A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fus. 24, 147\u2013164 (2015). https:\/\/doi.org\/10.1016\/j.inffus.2014.09.004","journal-title":"Inf. Fus."},{"key":"2759_CR52","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","volume":"82","author":"J Ma","year":"2017","unstructured":"Ma, J., Zhou, Z., Wang, B., Zong, H.: Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrar. Phys. Technol. 82, 8\u201317 (2017)","journal-title":"Infrar. Phys. Technol."},{"key":"2759_CR53","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1016\/j.inffus.2015.11.003","volume":"30","author":"Z Zhou","year":"2016","unstructured":"Zhou, Z., Wang, B., Li, S., Dong, M.: Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters. Inf. Fusion. 30, 15\u201326 (2016). https:\/\/doi.org\/10.1016\/j.inffus.2015.11.003","journal-title":"Inf. Fusion."},{"key":"2759_CR54","doi-asserted-by":"publisher","first-page":"308","DOI":"10.1049\/el:20000267","volume":"36","author":"C Xydeas","year":"2000","unstructured":"Xydeas, C., Petrovic, V.: Objective image fusion performance measure. Electron. Lett. 36, 308\u2013309 (2000)","journal-title":"Electron. Lett."},{"key":"2759_CR55","doi-asserted-by":"publisher","first-page":"1421","DOI":"10.1016\/j.imavis.2007.12.002","volume":"27","author":"Y Chen","year":"2009","unstructured":"Chen, Y., Blum, R.S.: A new automated quality assessment algorithm for image fusion. Image Vis. Comput. 27, 1421\u20131432 (2009). https:\/\/doi.org\/10.1016\/j.imavis.2007.12.002","journal-title":"Image Vis. Comput."},{"key":"2759_CR56","doi-asserted-by":"publisher","first-page":"193","DOI":"10.1016\/j.inffus.2005.10.001","volume":"8","author":"H Chen","year":"2007","unstructured":"Chen, H., Varshney, P.K.: A human perception inspired quality metric for image fusion based on regional information. Inf. Fusion. 8, 193\u2013207 (2007). https:\/\/doi.org\/10.1016\/j.inffus.2005.10.001","journal-title":"Inf. Fusion."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-022-02759-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-022-02759-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-022-02759-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,10]],"date-time":"2023-11-10T03:05:19Z","timestamp":1699585519000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-022-02759-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,1,16]]},"references-count":56,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2023,12]]}},"alternative-id":["2759"],"URL":"https:\/\/doi.org\/10.1007\/s00371-022-02759-w","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,1,16]]},"assertion":[{"value":"16 December 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 January 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there is no conflict of interest regarding the publication of the article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}