{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T18:23:01Z","timestamp":1773512581898,"version":"3.50.1"},"reference-count":49,"publisher":"MDPI AG","issue":"3","license":[{"start":{"date-parts":[[2023,1,24]],"date-time":"2023-01-24T00:00:00Z","timestamp":1674518400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"National Natural Science Foundation of China","award":["62175054"],"award-info":[{"award-number":["62175054"]}]},{"name":"National Natural Science Foundation of China","award":["61865005"],"award-info":[{"award-number":["61865005"]}]},{"name":"National Natural Science Foundation of China","award":["61762033"],"award-info":[{"award-number":["61762033"]}]},{"name":"National Natural Science Foundation of China","award":["620RC554"],"award-info":[{"award-number":["620RC554"]}]},{"name":"National Natural Science Foundation of China","award":["617079"],"award-info":[{"award-number":["617079"]}]},{"name":"National Natural Science Foundation of China","award":["2021-002"],"award-info":[{"award-number":["2021-002"]}]},{"name":"National Natural Science Foundation of China","award":["2020WNLOKF001"],"award-info":[{"award-number":["2020WNLOKF001"]}]},{"name":"National Natural Science Foundation of China","award":["2015BAH55F04"],"award-info":[{"award-number":["2015BAH55F04"]}]},{"name":"National Natural Science Foundation of China","award":["2015BAH55F01"],"award-info":[{"award-number":["2015BAH55F01"]}]},{"name":"National Natural Science Foundation of China","award":["ZDKJ2016015"],"award-info":[{"award-number":["ZDKJ2016015"]}]},{"name":"National Natural Science Foundation of China","award":["KYQD(ZR)1882"],"award-info":[{"award-number":["KYQD(ZR)1882"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["62175054"],"award-info":[{"award-number":["62175054"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["61865005"],"award-info":[{"award-number":["61865005"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["61762033"],"award-info":[{"award-number":["61762033"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["620RC554"],"award-info":[{"award-number":["620RC554"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["617079"],"award-info":[{"award-number":["617079"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["2021-002"],"award-info":[{"award-number":["2021-002"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["2020WNLOKF001"],"award-info":[{"award-number":["2020WNLOKF001"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["2015BAH55F04"],"award-info":[{"award-number":["2015BAH55F04"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["2015BAH55F01"],"award-info":[{"award-number":["2015BAH55F01"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["ZDKJ2016015"],"award-info":[{"award-number":["ZDKJ2016015"]}]},{"name":"Natural Science Foundation of Hainan Province","award":["KYQD(ZR)1882"],"award-info":[{"award-number":["KYQD(ZR)1882"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["62175054"],"award-info":[{"award-number":["62175054"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["61865005"],"award-info":[{"award-number":["61865005"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["61762033"],"award-info":[{"award-number":["61762033"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["620RC554"],"award-info":[{"award-number":["620RC554"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["617079"],"award-info":[{"award-number":["617079"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["2021-002"],"award-info":[{"award-number":["2021-002"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["2020WNLOKF001"],"award-info":[{"award-number":["2020WNLOKF001"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["2015BAH55F04"],"award-info":[{"award-number":["2015BAH55F04"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["2015BAH55F01"],"award-info":[{"award-number":["2015BAH55F01"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["ZDKJ2016015"],"award-info":[{"award-number":["ZDKJ2016015"]}]},{"name":"Major Science and Technology Project of Haikou City","award":["KYQD(ZR)1882"],"award-info":[{"award-number":["KYQD(ZR)1882"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["62175054"],"award-info":[{"award-number":["62175054"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["61865005"],"award-info":[{"award-number":["61865005"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["61762033"],"award-info":[{"award-number":["61762033"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["620RC554"],"award-info":[{"award-number":["620RC554"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["617079"],"award-info":[{"award-number":["617079"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["2021-002"],"award-info":[{"award-number":["2021-002"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["2020WNLOKF001"],"award-info":[{"award-number":["2020WNLOKF001"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["2015BAH55F04"],"award-info":[{"award-number":["2015BAH55F04"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["2015BAH55F01"],"award-info":[{"award-number":["2015BAH55F01"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["ZDKJ2016015"],"award-info":[{"award-number":["ZDKJ2016015"]}]},{"name":"Open Project Program of Wuhan National Laboratory for Optoelectronics","award":["KYQD(ZR)1882"],"award-info":[{"award-number":["KYQD(ZR)1882"]}]},{"name":"National Key Technology Support Program","award":["62175054"],"award-info":[{"award-number":["62175054"]}]},{"name":"National Key Technology Support Program","award":["61865005"],"award-info":[{"award-number":["61865005"]}]},{"name":"National Key Technology Support Program","award":["61762033"],"award-info":[{"award-number":["61762033"]}]},{"name":"National Key Technology Support Program","award":["620RC554"],"award-info":[{"award-number":["620RC554"]}]},{"name":"National Key Technology Support Program","award":["617079"],"award-info":[{"award-number":["617079"]}]},{"name":"National Key Technology Support Program","award":["2021-002"],"award-info":[{"award-number":["2021-002"]}]},{"name":"National Key Technology Support Program","award":["2020WNLOKF001"],"award-info":[{"award-number":["2020WNLOKF001"]}]},{"name":"National Key Technology Support Program","award":["2015BAH55F04"],"award-info":[{"award-number":["2015BAH55F04"]}]},{"name":"National Key Technology Support Program","award":["2015BAH55F01"],"award-info":[{"award-number":["2015BAH55F01"]}]},{"name":"National Key Technology Support Program","award":["ZDKJ2016015"],"award-info":[{"award-number":["ZDKJ2016015"]}]},{"name":"National Key Technology Support Program","award":["KYQD(ZR)1882"],"award-info":[{"award-number":["KYQD(ZR)1882"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["62175054"],"award-info":[{"award-number":["62175054"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["61865005"],"award-info":[{"award-number":["61865005"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["61762033"],"award-info":[{"award-number":["61762033"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["620RC554"],"award-info":[{"award-number":["620RC554"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["617079"],"award-info":[{"award-number":["617079"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["2021-002"],"award-info":[{"award-number":["2021-002"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["2020WNLOKF001"],"award-info":[{"award-number":["2020WNLOKF001"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["2015BAH55F04"],"award-info":[{"award-number":["2015BAH55F04"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["2015BAH55F01"],"award-info":[{"award-number":["2015BAH55F01"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["ZDKJ2016015"],"award-info":[{"award-number":["ZDKJ2016015"]}]},{"name":"Major Science and Technology Project of Hainan Province","award":["KYQD(ZR)1882"],"award-info":[{"award-number":["KYQD(ZR)1882"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["62175054"],"award-info":[{"award-number":["62175054"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["61865005"],"award-info":[{"award-number":["61865005"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["61762033"],"award-info":[{"award-number":["61762033"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["620RC554"],"award-info":[{"award-number":["620RC554"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["617079"],"award-info":[{"award-number":["617079"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["2021-002"],"award-info":[{"award-number":["2021-002"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["2020WNLOKF001"],"award-info":[{"award-number":["2020WNLOKF001"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["2015BAH55F04"],"award-info":[{"award-number":["2015BAH55F04"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["2015BAH55F01"],"award-info":[{"award-number":["2015BAH55F01"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["ZDKJ2016015"],"award-info":[{"award-number":["ZDKJ2016015"]}]},{"name":"Scientific Research Staring Foundation of Hainan University","award":["KYQD(ZR)1882"],"award-info":[{"award-number":["KYQD(ZR)1882"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>The aim of infrared (IR) and visible image fusion is to generate a more informative image for human observation or some other computer vision tasks. The activity-level measurement and weight assignment are two key parts in image fusion. In this paper, we propose a novel IR and visible fusion method based on the principal component analysis network (PCANet) and an image pyramid. Firstly, we use the lightweight deep learning network, a PCANet, to obtain the activity-level measurement and weight assignment of IR and visible images. The activity-level measurement obtained by the PCANet has a stronger representation ability for focusing on IR target perception and visible detail description. Secondly, the weights and the source images are decomposed into multiple scales by the image pyramid, and the weighted-average fusion rule is applied at each scale. Finally, the fused image is obtained by reconstruction. The effectiveness of the proposed algorithm was verified by two datasets with more than eighty pairs of test images in total. Compared with nineteen representative methods, the experimental results demonstrate that the proposed method can achieve the state-of-the-art results in both visual quality and objective evaluation metrics.<\/jats:p>","DOI":"10.3390\/rs15030685","type":"journal-article","created":{"date-parts":[[2023,1,25]],"date-time":"2023-01-25T04:22:16Z","timestamp":1674620536000},"page":"685","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Infrared and Visible Image Fusion Method Based on a Principal Component Analysis Network and Image Pyramid"],"prefix":"10.3390","volume":"15","author":[{"given":"Shengshi","family":"Li","sequence":"first","affiliation":[{"name":"School of Information and Communication Engineering, Hainan University, Haikou 570228, China"}]},{"given":"Yonghua","family":"Zou","sequence":"additional","affiliation":[{"name":"School of Information and Communication Engineering, Hainan University, Haikou 570228, China"},{"name":"State Key Laboratory of Marine Resource Utilization in South China Sea, Hainan University, Haikou 570228, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5458-9509","authenticated-orcid":false,"given":"Guanjun","family":"Wang","sequence":"additional","affiliation":[{"name":"School of Information and Communication Engineering, Hainan University, Haikou 570228, China"},{"name":"State Key Laboratory of Marine Resource Utilization in South China Sea, Hainan University, Haikou 570228, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1567-1398","authenticated-orcid":false,"given":"Cong","family":"Lin","sequence":"additional","affiliation":[{"name":"School of Information and Communication Engineering, Hainan University, Haikou 570228, China"}]}],"member":"1968","published-online":{"date-parts":[[2023,1,24]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Qi, B., Jin, L., Li, G., Zhang, Y., Li, Q., Bi, G., and Wang, W. (2022). Infrared and Visible Image Fusion Based on Co-Occurrence Analysis Shearlet Transform. Remote Sens., 14.","DOI":"10.3390\/rs14020283"},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Gao, X., Shi, Y., Zhu, Q., Fu, Q., and Wu, Y. (2022). Infrared and Visible Image Fusion with Deep Neural Network in Enhanced Flight Vision System. Remote Sens., 14.","DOI":"10.3390\/rs14122789"},{"key":"ref_3","unstructured":"Burt, P.J., and Adelson, E.H. (1987). Readings in Computer Vision, Elsevier."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"479","DOI":"10.14429\/dsj.61.705","article-title":"Image fusion technique using multi-resolution singular value decomposition","volume":"61","author":"Naidu","year":"2011","journal-title":"Defence Sci. J."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"235","DOI":"10.1006\/gmip.1995.1022","article-title":"Multisensor image fusion using the wavelet transform","volume":"57","author":"Li","year":"1995","journal-title":"Gr. Models Image Process."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"119","DOI":"10.1016\/j.inffus.2005.09.006","article-title":"Pixel-and region-based image fusion with complex wavelets","volume":"8","author":"Lewis","year":"2007","journal-title":"Inf. Fusion"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"143","DOI":"10.1016\/j.inffus.2006.02.001","article-title":"Remote sensing image fusion using the curvelet transform","volume":"8","author":"Nencini","year":"2007","journal-title":"Inf. Fusion"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","article-title":"Infrared and visible image fusion based on target-enhanced multiscale transform decomposition","volume":"508","author":"Chen","year":"2020","journal-title":"Inf. Sci."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"255","DOI":"10.1016\/j.ijleo.2016.09.126","article-title":"Texture clear multi-modal image fusion with joint sparsity model","volume":"130","author":"Gao","year":"2017","journal-title":"Optik"},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"057006","DOI":"10.1117\/1.OE.52.5.057006","article-title":"Dictionary learning method for joint sparse representation-based image fusion","volume":"52","author":"Zhang","year":"2013","journal-title":"Opt. Eng."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"94","DOI":"10.1016\/j.infrared.2017.04.018","article-title":"Infrared and visible image fusion method based on saliency detection in sparse domain","volume":"83","author":"Liu","year":"2017","journal-title":"Infrared Phys. Technol."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","article-title":"Infrared and visible image fusion based on visual saliency map and weighted least square optimization","volume":"82","author":"Ma","year":"2017","journal-title":"Infrared Phys. Technol."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"824","DOI":"10.1109\/TCI.2021.3100986","article-title":"Classification saliency-based rule for visible and infrared image fusion","volume":"7","author":"Xu","year":"2021","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","article-title":"Multi-focus image fusion with a deep convolutional neural network","volume":"36","author":"Liu","year":"2017","journal-title":"Inf. Fusion"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"1850018","DOI":"10.1142\/S0219691318500182","article-title":"Infrared and visible image fusion with convolutional neural networks","volume":"16","author":"Liu","year":"2018","journal-title":"Int. J. Wavel. Multiresolut. Inf. Process."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Liu, Y., Chen, X., Cheng, J., and Peng, H. (2017, January 10\u201313). A medical image fusion method based on convolutional neural networks. Proceedings of the 2017 20th International Conference on Information Fusion (Fusion), Xi\u2019an, China.","DOI":"10.23919\/ICIF.2017.8009769"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Li, H., Wu, X.J., and Kittler, J. (2018, January 20\u201324). Infrared and visible image fusion using a deep learning framework. Proceedings of the 2018 24th international conference on pattern recognition (ICPR), Beijing, China.","DOI":"10.1109\/ICPR.2018.8546006"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"103039","DOI":"10.1016\/j.infrared.2019.103039","article-title":"Infrared and visible image fusion with ResNet and zero-phase component analysis","volume":"102","author":"Li","year":"2019","journal-title":"Infrared Phys. Technol."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","article-title":"FusionGAN: A generative adversarial network for infrared and visible image fusion","volume":"48","author":"Ma","year":"2019","journal-title":"Inf. Fusion"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","article-title":"DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion","volume":"29","author":"Ma","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"ref_21","first-page":"1","article-title":"GANMcC: A generative adversarial network with multiclassification constraints for infrared and visible image fusion","volume":"70","author":"Ma","year":"2020","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"5017","DOI":"10.1109\/TIP.2015.2475625","article-title":"PCANet: A simple deep learning baseline for image classification?","volume":"24","author":"Chan","year":"2015","journal-title":"IEEE Trans. Image Process."},{"key":"ref_23","unstructured":"Mertens, T., Kautz, J., and Van Reeth, F. (November, January 29). Exposure fusion. Proceedings of the 15th Pacific Conference on Computer Graphics and Applications (PG\u201907), Seoul, Republic of Korea."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"259","DOI":"10.1016\/S1566-2535(03)00046-0","article-title":"A general framework for multiresolution image fusion: From pixels to regions","volume":"4","author":"Piella","year":"2003","journal-title":"Inf. Fusion"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"11603","DOI":"10.1007\/s11042-015-2698-y","article-title":"Human fall detection in surveillance video based on PCANet","volume":"75","author":"Wang","year":"2016","journal-title":"Multimed. Tools Appl."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"1792","DOI":"10.1109\/LGRS.2016.2611001","article-title":"Automatic change detection in synthetic aperture radar images based on PCANet","volume":"13","author":"Gao","year":"2016","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Song, X., and Wu, X.J. (2018, January 20). Multi-focus image fusion with PCA filters of PCANet. Proceedings of the IAPR Workshop on Multimodal Pattern Recognition of Social Signals in Human\u2013Computer Interaction, Beijing, China.","DOI":"10.1007\/978-3-030-20984-1_1"},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"22","DOI":"10.1016\/j.compbiomed.2018.08.003","article-title":"Automatic recognition of arrhythmia based on principal component analysis network and linear support vector machine","volume":"101","author":"Yang","year":"2018","journal-title":"Comput. Biol. Med."},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"171570","DOI":"10.1109\/ACCESS.2019.2955555","article-title":"Automated detection of myocardial infarction using a gramian angular field and principal component analysis network","volume":"7","author":"Zhang","year":"2019","journal-title":"IEEE Access"},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"1397","DOI":"10.1109\/TPAMI.2012.213","article-title":"Guided image filtering","volume":"35","author":"He","year":"2012","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_31","unstructured":"Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., and Zitnick, C.L. (2014). European Conference on Computer Vision, Springer."},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"2864","DOI":"10.1109\/TIP.2013.2244222","article-title":"Image fusion with guided filtering","volume":"22","author":"Li","year":"2013","journal-title":"IEEE Trans. Image Process."},{"key":"ref_33","unstructured":"Toet, A. (2022, September 21). TNO Image Fusion Dataset. Available online: https:\/\/figshare.com\/articles\/TN_Image_Fusion_Dataset\/1008029."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Xu, H., Ma, J., Le, Z., Jiang, J., and Guo, X. (2020, January 7\u201312). Fusiondn: A unified densely connected network for image fusion. Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA.","DOI":"10.1609\/aaai.v34i07.6936"},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"156","DOI":"10.1016\/j.inffus.2006.09.001","article-title":"A novel similarity based quality metric for image fusion","volume":"9","author":"Yang","year":"2008","journal-title":"Inf. Fusion"},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"308","DOI":"10.1049\/el:20000267","article-title":"Objective image fusion performance measure","volume":"36","author":"Xydeas","year":"2000","journal-title":"Electron. Lett."},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image quality assessment: From error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Haghighat, M., and Razian, M.A. (2014, January 15\u201317). Fast-FMI: Non-reference image fusion metric. Proceedings of the 2014 IEEE 8th International Conference on Application of Information and Communication Technologies (AICT), Astana, Kazakhstan.","DOI":"10.1109\/ICAICT.2014.7036000"},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"1125","DOI":"10.1007\/s11760-012-0361-x","article-title":"Multifocus and multispectral image fusion based on pixel significance using discrete cosine harmonic wavelet transform","volume":"7","year":"2013","journal-title":"Signal Image Video Process."},{"key":"ref_40","unstructured":"Piella, G., and Heijmans, H. (2003, January 14\u201317). A new quality metric for image fusion. Proceedings of the 2003 International Conference on Image Processing (Cat. No. 03CH37429), Barcelona, Spain."},{"key":"ref_41","first-page":"1433","article-title":"Performance assessment of combinative pixel-level image fusion based on an absolute feature measurement","volume":"3","author":"Zhao","year":"2007","journal-title":"Int. J. Innov. Comput. Inf. Control"},{"key":"ref_42","doi-asserted-by":"crossref","first-page":"193","DOI":"10.1016\/j.inffus.2005.10.001","article-title":"A human perception inspired quality metric for image fusion based on regional information","volume":"8","author":"Chen","year":"2007","journal-title":"Inf. Fusion"},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"1421","DOI":"10.1016\/j.imavis.2007.12.002","article-title":"A new automated quality assessment algorithm for image fusion","volume":"27","author":"Chen","year":"2009","journal-title":"Image Vis. Comput."},{"key":"ref_44","doi-asserted-by":"crossref","first-page":"3064","DOI":"10.1364\/AO.58.003064","article-title":"Infrared and visible image perceptive fusion through multi-level Gaussian curvature filtering image decomposition","volume":"58","author":"Tan","year":"2019","journal-title":"Appl. Opt."},{"key":"ref_45","doi-asserted-by":"crossref","unstructured":"Zhang, H., Xu, H., Xiao, Y., Guo, X., and Ma, J. (2020, January 7\u201312). Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA.","DOI":"10.1609\/aaai.v34i07.6975"},{"key":"ref_46","doi-asserted-by":"crossref","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","article-title":"RFN-Nest: An end-to-end residual fusion network for infrared and visible images","volume":"73","author":"Li","year":"2021","journal-title":"Inf. Fusion"},{"key":"ref_47","first-page":"1","article-title":"DRF: Disentangled representation for visible and infrared image fusion","volume":"70","author":"Xu","year":"2021","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"ref_48","doi-asserted-by":"crossref","first-page":"100","DOI":"10.1016\/j.inffus.2016.02.001","article-title":"Infrared and visible image fusion via gradient transfer and total variation minimization","volume":"31","author":"Ma","year":"2016","journal-title":"Inf. Fusion"},{"key":"ref_49","doi-asserted-by":"crossref","unstructured":"Du, Q., Xu, H., Ma, Y., Huang, J., and Fan, F. (2018). Fusing infrared and visible images of different resolutions via total variation model. Sensors, 18.","DOI":"10.3390\/s18113827"}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/3\/685\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T18:14:32Z","timestamp":1760120072000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/3\/685"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,1,24]]},"references-count":49,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2023,2]]}},"alternative-id":["rs15030685"],"URL":"https:\/\/doi.org\/10.3390\/rs15030685","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,1,24]]}}}