{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,16]],"date-time":"2026-01-16T07:45:43Z","timestamp":1768549543243,"version":"3.49.0"},"reference-count":35,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2024,11,5]],"date-time":"2024-11-05T00:00:00Z","timestamp":1730764800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,5]],"date-time":"2024-11-05T00:00:00Z","timestamp":1730764800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U23A20384"],"award-info":[{"award-number":["U23A20384"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["No. 3132023523"],"award-info":[{"award-number":["No. 3132023523"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1007\/s11554-024-01570-1","type":"journal-article","created":{"date-parts":[[2024,11,5]],"date-time":"2024-11-05T06:07:43Z","timestamp":1730786863000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["FLFusionSR: a fast and lightweight fusion and super-resolution network for infrared and visible images on edge devices"],"prefix":"10.1007","volume":"21","author":[{"given":"Weimin","family":"Xue","sequence":"first","affiliation":[]},{"given":"Yisha","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Guojian","family":"He","sequence":"additional","affiliation":[]},{"given":"Fei","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yan","family":"Zhuang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,5]]},"reference":[{"issue":"6","key":"1570_CR1","doi-asserted-by":"publisher","first-page":"2479","DOI":"10.1007\/s11554-020-01068-6","volume":"18","author":"M Awad","year":"2021","unstructured":"Awad, M., Elliethy, A., Aly, H.A.: Real-time visible and near-infrared video fusion: architecture and implementation. J. Real-Time Image Proc. 18(6), 2479\u20132493 (2021)","journal-title":"J. Real-Time Image Proc."},{"issue":"6","key":"1570_CR2","doi-asserted-by":"publisher","first-page":"2331","DOI":"10.1007\/s11554-021-01111-0","volume":"18","author":"J Li","year":"2021","unstructured":"Li, J., Peng, Y., Jiang, T.: Embedded real-time infrared and visible image fusion for uav surveillance. J. Real-Time Image Proc. 18(6), 2331\u20132345 (2021)","journal-title":"J. Real-Time Image Proc."},{"key":"1570_CR3","doi-asserted-by":"crossref","unstructured":"Qi, J., Abera, D.E., Fanose, M.N., Wang, L., Cheng, J.: A deep learning and image enhancement based pipeline for infrared and visible image fusion. Neurocomputing, 127353 (2024)","DOI":"10.1016\/j.neucom.2024.127353"},{"issue":"6","key":"1570_CR4","doi-asserted-by":"publisher","first-page":"117","DOI":"10.1007\/s11554-023-01367-8","volume":"20","author":"Q Zhao","year":"2023","unstructured":"Zhao, Q., Ji, T., Liang, S., Yu, W., Yan, C.: Real-time power line segmentation detection based on multi-attention with strong semantic feature extractor. J. Real-Time Image Proc. 20(6), 117 (2023)","journal-title":"J. Real-Time Image Proc."},{"key":"1570_CR5","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.102039","volume":"102","author":"X Wang","year":"2024","unstructured":"Wang, X., Guan, Z., Qian, W., Cao, J., Liang, S., Yan, J.: CS2Fusion: contrastive learning for self-supervised infrared and visible image fusion by estimating feature compensation map. Inf. Fusion 102, 102039 (2024)","journal-title":"Inf. Fusion"},{"issue":"1","key":"1570_CR6","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1570_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2022.105253","volume":"144","author":"MA Azam","year":"2022","unstructured":"Azam, M.A., Khan, K.B., Salahuddin, S., Rehman, E., Khan, S.A., Khan, M.A., Kadry, S., Gandomi, A.H.: A review on multimodal medical image fusion: compendious analysis of medical modalities, multimodal databases, fusion techniques and quality metrics. Comput. Biol. Med. 144, 105253 (2022)","journal-title":"Comput. Biol. Med."},{"issue":"3","key":"1570_CR8","doi-asserted-by":"publisher","first-page":"771","DOI":"10.3390\/rs14030771","volume":"14","author":"F Xu","year":"2022","unstructured":"Xu, F., Liu, J., Song, Y., Sun, H., Wang, X.: Multi-exposure image fusion techniques: a comprehensive review. Remote Sens. 14(3), 771 (2022)","journal-title":"Remote Sens."},{"issue":"2","key":"1570_CR9","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11554-024-01415-x","volume":"21","author":"Y Lei","year":"2024","unstructured":"Lei, Y., Wang, X., An, A., Guan, H.: Deeplab-yolo: a method for detecting hot-spot defects in infrared image pv panels by combining segmentation and detection. J. Real-Time Image Proc. 21(2), 1\u201311 (2024)","journal-title":"J. Real-Time Image Proc."},{"issue":"3","key":"1570_CR10","doi-asserted-by":"publisher","first-page":"1807","DOI":"10.1109\/JSEN.2021.3139124","volume":"22","author":"AN Wilson","year":"2022","unstructured":"Wilson, A.N., Kumar, A., Jha, A., Cenkeramaddi, L.R.: Embedded sensors, communication technologies, computing platforms and machine learning for UAVs: a review. IEEE Sens. J. 22(3), 1807\u20131826 (2022)","journal-title":"IEEE Sens. J."},{"key":"1570_CR11","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","volume":"76","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Xu, H., Tian, X., Jiang, J., Ma, J.: Image fusion meets deep learning: a survey and perspective. Inf. Fusion 76, 323\u2013336 (2021)","journal-title":"Inf. Fusion"},{"issue":"8","key":"1570_CR12","doi-asserted-by":"publisher","first-page":"10535","DOI":"10.1109\/TPAMI.2023.3261282","volume":"45","author":"X Zhang","year":"2023","unstructured":"Zhang, X., Demiris, Y.: Visible and infrared image fusion using deep learning. IEEE Trans. Pattern Anal. Mach. Intell. 45(8), 10535\u201310554 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1570_CR13","first-page":"1","volume":"71","author":"W Xiao","year":"2022","unstructured":"Xiao, W., Zhang, Y., Wang, H., Li, F., Jin, H.: Heterogeneous knowledge distillation for simultaneous infrared-visible image fusion and super-resolution. IEEE Trans. Instrum. Meas. 71, 1\u201315 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"1570_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2022.104383","volume":"127","author":"W Xue","year":"2022","unstructured":"Xue, W., Wang, A., Zhao, L.: FLFuse-Net: a fast and lightweight infrared and visible image fusion network via feature flow and edge compensation for salient information. Infrared Phys. Technol. 127, 104383 (2022)","journal-title":"Infrared Phys. Technol."},{"issue":"10","key":"1570_CR15","doi-asserted-by":"publisher","first-page":"2761","DOI":"10.1007\/s11263-021-01501-8","volume":"129","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Ma, J.: SDNet: a versatile squeeze-and-decomposition network for real-time image fusion. Int. J. Comput. Vis. 129(10), 2761\u20132785 (2021)","journal-title":"Int. J. Comput. Vis."},{"key":"1570_CR16","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 82, 28\u201342 (2022)","journal-title":"Inf. Fusion"},{"issue":"5","key":"1570_CR17","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.J.: DenseFuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"1570_CR18","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: FusionGAN: a generative adversarial network for infrared and visible image fusion. Inf. Fusion 48, 11\u201326 (2019)","journal-title":"Inf. Fusion"},{"key":"1570_CR19","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1016\/j.inffus.2019.07.005","volume":"54","author":"J Ma","year":"2020","unstructured":"Ma, J., Liang, P., Yu, W., Chen, C., Guo, X., Wu, J., Jiang, J.: Infrared and visible image fusion via detail preserving adversarial learning. Inf. Fusion 54, 85\u201398 (2020)","journal-title":"Inf. Fusion"},{"key":"1570_CR20","doi-asserted-by":"crossref","unstructured":"Zhong, J., Yang, B., Li, Y., Zhong, F., Chen, Z.: Image fusion and super-resolution with convolutional neural network. In: Chinese Conference on Pattern Recognition, pp. 78\u201388. Springer (2016)","DOI":"10.1007\/978-981-10-3005-5_7"},{"issue":"3","key":"1570_CR21","doi-asserted-by":"publisher","first-page":"229","DOI":"10.1016\/j.inffus.2012.01.008","volume":"14","author":"H Yin","year":"2013","unstructured":"Yin, H., Li, S., Fang, L.: Simultaneous image fusion and super-resolution using sparse representation. Inf. Fusion 14(3), 229\u2013240 (2013)","journal-title":"Inf. Fusion"},{"key":"1570_CR22","doi-asserted-by":"publisher","first-page":"138","DOI":"10.1016\/j.neucom.2015.06.035","volume":"171","author":"H Li","year":"2016","unstructured":"Li, H., Yu, Z., Mao, C.: Fractional differential and variational method for image fusion and super-resolution. Neurocomputing 171, 138\u2013148 (2016)","journal-title":"Neurocomputing"},{"key":"1570_CR23","doi-asserted-by":"publisher","first-page":"174092","DOI":"10.1109\/ACCESS.2019.2957165","volume":"7","author":"M Xie","year":"2019","unstructured":"Xie, M., Zhou, Z., Zhang, Y.: Joint framework for image fusion and super-resolution via multicomponent analysis and residual compensation. IEEE Access 7, 174092\u2013174107 (2019)","journal-title":"IEEE Access"},{"key":"1570_CR24","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.C.: Mobilenetv2: Inverted residuals and linear bottlenecks. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1570_CR25","doi-asserted-by":"crossref","unstructured":"Shi, W., Caballero, J., Husz\u00e1r, F., Totz, J., Aitken, A.P., Bishop, R., Rueckert, D., Wang, Z.: Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1874\u20131883 (2016)","DOI":"10.1109\/CVPR.2016.207"},{"key":"1570_CR26","unstructured":"Liu, S., Davison, A.J., Johns, E.: Self-supervised generalisation with meta auxiliary learning. In: Neural Information Processing Systems, pp. 1679\u20131689 (2019)"},{"issue":"1","key":"1570_CR27","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1109\/TCI.2016.2644865","volume":"3","author":"H Zhao","year":"2017","unstructured":"Zhao, H., Gallo, O., Frosio, I., Kautz, J.: Loss functions for image restoration with neural networks. IEEE Trans. Comput. Imaging 3(1), 47\u201357 (2017)","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"1570_CR28","doi-asserted-by":"crossref","unstructured":"Brown, M., S\u00fcsstrunk, S.: Multi-spectral sift for scene category recognition. In: CVPR 2011, pp. 177\u2013184. IEEE (2011)","DOI":"10.1109\/CVPR.2011.5995637"},{"issue":"1","key":"1570_CR29","doi-asserted-by":"publisher","DOI":"10.1117\/1.OE.51.1.010901","volume":"51","author":"A Toet","year":"2012","unstructured":"Toet, A., Hogervorst, M.A.: Progress in color night vision. Opt. Eng. 51(1), 010901 (2012). https:\/\/doi.org\/10.1117\/1.OE.51.1.010901","journal-title":"Opt. Eng."},{"issue":"5","key":"1570_CR30","doi-asserted-by":"publisher","first-page":"744","DOI":"10.1016\/j.compeleceng.2011.07.012","volume":"37","author":"MBA Haghighat","year":"2011","unstructured":"Haghighat, M.B.A., Aghagolzadeh, A., Seyedarabi, H.: A non-reference image fusion metric based on mutual information of image features. Comput. Electr. Eng. 37(5), 744\u2013756 (2011)","journal-title":"Comput. Electr. Eng."},{"issue":"6","key":"1570_CR31","doi-asserted-by":"publisher","first-page":"1125","DOI":"10.1007\/s11760-012-0361-x","volume":"7","author":"BS Kumar","year":"2013","unstructured":"Kumar, B.S.: Multifocus and multispectral image fusion based on pixel significance using discrete cosine harmonic wavelet transform. SIViP 7(6), 1125\u20131143 (2013)","journal-title":"SIViP"},{"issue":"4","key":"1570_CR32","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P., et al.: Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"1570_CR33","unstructured":"Wang, Z., Simoncelli, E.P., Bovik, A.C.: Multiscale structural similarity for image quality assessment. In: The Thrity-Seventh Asilomar Conference on Signals, Systems and Computers, 2003, vol.\u00a02, pp. 1398\u20131402. IEEE (2003)"},{"key":"1570_CR34","doi-asserted-by":"publisher","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","volume":"29","author":"J Ma","year":"2020","unstructured":"Ma, J., Xu, H., Jiang, J., Mei, X., Zhang, X.P.: DDcGAN: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 29, 4980\u20134995 (2020)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"1570_CR35","doi-asserted-by":"publisher","first-page":"1934","DOI":"10.1109\/TPAMI.2022.3167175","volume":"45","author":"SW Zamir","year":"2023","unstructured":"Zamir, S.W., Arora, A., Khan, S., Hayat, M., Khan, F.S., Yang, M.H., Shao, L.: Learning enriched features for fast image restoration and enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 45(2), 1934\u20131948 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01570-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-024-01570-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-024-01570-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T13:17:33Z","timestamp":1732799853000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-024-01570-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,5]]},"references-count":35,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["1570"],"URL":"https:\/\/doi.org\/10.1007\/s11554-024-01570-1","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"value":"1861-8200","type":"print"},{"value":"1861-8219","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,5]]},"assertion":[{"value":"25 May 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 October 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 November 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"192"}}