{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,8]],"date-time":"2025-12-08T22:36:05Z","timestamp":1765233365478,"version":"3.37.3"},"reference-count":67,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2022,8,25]],"date-time":"2022-08-25T00:00:00Z","timestamp":1661385600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,8,25]],"date-time":"2022-08-25T00:00:00Z","timestamp":1661385600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61733002"],"award-info":[{"award-number":["61733002"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100018617","name":"LiaoNing Revitalization Talents Program","doi-asserted-by":"crossref","award":["XLYC1807088"],"award-info":[{"award-number":["XLYC1807088"]}],"id":[{"id":"10.13039\/501100018617","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61922019","62027826"],"award-info":[{"award-number":["61922019","62027826"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2023,10]]},"DOI":"10.1007\/s00371-022-02633-9","type":"journal-article","created":{"date-parts":[[2022,8,25]],"date-time":"2022-08-25T12:03:07Z","timestamp":1661428987000},"page":"4869-4886","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["A unified image fusion framework with flexible bilevel paradigm integration"],"prefix":"10.1007","volume":"39","author":[{"given":"Jinyuan","family":"Liu","sequence":"first","affiliation":[]},{"given":"Zhiying","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Guanyao","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Risheng","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Fan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,8,25]]},"reference":[{"key":"2633_CR1","doi-asserted-by":"crossref","unstructured":"Viola P, Jones M, et al.: Rapid object detection using a boosted cascade of simple features. In: CVPR, vol 1, pp 511\u2013518 (2001)","DOI":"10.1109\/CVPR.2001.990517"},{"issue":"9","key":"2633_CR2","doi-asserted-by":"crossref","first-page":"1627","DOI":"10.1109\/TPAMI.2009.167","volume":"32","author":"PF Felzenszwalb","year":"2010","unstructured":"Felzenszwalb, P.F., Girshick, R.B., McAllester, D., Ramanan, D.: Object detection with discriminatively trained part-based models. IEEE TPAMI 32(9), 1627\u20131645 (2010)","journal-title":"IEEE TPAMI"},{"key":"2633_CR3","doi-asserted-by":"crossref","unstructured":"Dikmen, M., Akbas, E., Huang, T.S., Ahuja, N.: Pedestrian recognition with a learned metric. In: ACCV, pp. 501\u2013512. Springer, Berlin (2010)","DOI":"10.1007\/978-3-642-19282-1_40"},{"key":"2633_CR4","doi-asserted-by":"crossref","unstructured":"Gray, D., Tao, H.: Viewpoint invariant pedestrian recognition with an ensemble of localized features. In: ECCV, pp. 262\u2013275. Springer, Berlin (2008)","DOI":"10.1007\/978-3-540-88682-2_21"},{"key":"2633_CR5","doi-asserted-by":"crossref","unstructured":"Duan, Z., Lan, J., Xu, T., Ni, B., Zhuang, L., Yang, X.: Pedestrian detection via bi-directional multi-scale analysis. In: ACM MM, pp. 1023\u20131031. ACM, New York (2017)","DOI":"10.1145\/3123266.3123356"},{"key":"2633_CR6","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: CVPR, pp. 3431\u20133440 (2015)","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"2633_CR7","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: CVPR, pp. 580\u2013587 (2014)","DOI":"10.1109\/CVPR.2014.81"},{"key":"2633_CR8","doi-asserted-by":"crossref","unstructured":"Pu, M., Huang, Y., Guan, Q., Zou, Q.: Graphnet: learning image pseudo annotations for weakly-supervised semantic segmentation. In: ACM MM, pp. 483\u2013491. ACM, New York (2018)","DOI":"10.1145\/3240508.3240542"},{"issue":"5","key":"2633_CR9","doi-asserted-by":"crossref","first-page":"1014","DOI":"10.1109\/TMM.2013.2244870","volume":"15","author":"G Bhatnagar","year":"2013","unstructured":"Bhatnagar, G., Wu, Q.M.J., Liu, Z.: Directive contrast based multimodal medical image fusion in NSCT domain. IEEE Trans. Multimedia 15(5), 1014\u20131024 (2013)","journal-title":"IEEE Trans. Multimedia"},{"issue":"7","key":"2633_CR10","doi-asserted-by":"crossref","first-page":"2864","DOI":"10.1109\/TIP.2013.2244222","volume":"22","author":"S Li","year":"2013","unstructured":"Li, S., Kang, X., Jianwen, H.: Image fusion with guided filtering. IEEE Trans. Image Process. 22(7), 2864\u20132875 (2013)","journal-title":"IEEE Trans. Image Process."},{"issue":"6","key":"2633_CR11","doi-asserted-by":"crossref","first-page":"123","DOI":"10.1109\/MSP.2005.1550194","volume":"22","author":"IW Selesnick","year":"2005","unstructured":"Selesnick, I.W., Baraniuk, R.G., Kingsbury, N.G.: The dual-tree complex wavelet transform. IEEE Signal Process. Mag. 22(6), 123\u2013151 (2005)","journal-title":"IEEE Signal Process. Mag."},{"issue":"9","key":"2633_CR12","doi-asserted-by":"crossref","first-page":"929","DOI":"10.1016\/S0167-8655(01)00047-2","volume":"22","author":"Z Liu","year":"2001","unstructured":"Liu, Z., Tsukada, K., Hanasaki, K., Ho, Y.-K., Dai, Y.P.: Image fusion by using steerable pyramid. Pattern Recogn. Lett. 22(9), 929\u2013939 (2001)","journal-title":"Pattern Recogn. Lett."},{"key":"2633_CR13","doi-asserted-by":"crossref","first-page":"94","DOI":"10.1016\/j.infrared.2013.07.010","volume":"61","author":"J Adu","year":"2013","unstructured":"Adu, J., Gan, J., Wang, Y., Huang, J.: Image fusion based on nonsubsampled contourlet transform for infrared and visible light image. Infrared Phys. Technol. 61, 94\u2013100 (2013)","journal-title":"Infrared Phys. Technol."},{"issue":"6583","key":"2633_CR14","doi-asserted-by":"crossref","first-page":"607","DOI":"10.1038\/381607a0","volume":"381","author":"A Bruno Olshausen","year":"1996","unstructured":"Bruno Olshausen, A., David, F.J.: Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature 381(6583), 607 (1996)","journal-title":"Nature"},{"key":"2633_CR15","doi-asserted-by":"crossref","first-page":"397","DOI":"10.1016\/j.infrared.2014.09.007","volume":"67","author":"L Xiaoqi","year":"2014","unstructured":"Xiaoqi, L., Zhang, B., Zhao, Y., Liu, H., Pei, H.: The infrared and visible image fusion algorithm based on target separation and sparse representation. Infrared Phys. Technol. 67, 397\u2013407 (2014)","journal-title":"Infrared Phys. Technol."},{"key":"2633_CR16","doi-asserted-by":"crossref","first-page":"57","DOI":"10.1016\/j.inffus.2017.05.006","volume":"40","author":"Q Zhang","year":"2018","unstructured":"Zhang, Q., Liu, Y., Blum, R.S., Jungong, H., Dacheng, T.: Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images. A review. Inf. Fus. 40, 57\u201375 (2018)","journal-title":"Inf. Fus."},{"key":"2633_CR17","doi-asserted-by":"crossref","first-page":"516","DOI":"10.1016\/j.ins.2017.09.010","volume":"432","author":"Z Zhu","year":"2018","unstructured":"Zhu, Z., Yin, H., Chai, Y., Li, Y., Qi, G.: A novel multi-modality image fusion method based on image decomposition and sparse representation. Inf. Sci. 432, 516\u2013529 (2018)","journal-title":"Inf. Sci."},{"issue":"3","key":"2633_CR18","doi-asserted-by":"crossref","first-page":"485","DOI":"10.1109\/LSP.2019.2895749","volume":"26","author":"Y Liu","year":"2019","unstructured":"Liu, Y., Chen, X., Ward, R.K., Wang, J.Z.: Medical image fusion via convolutional sparsity based morphological component analysis. IEEE Signal Process. Lett. 26(3), 485\u2013489 (2019)","journal-title":"IEEE Signal Process. Lett."},{"issue":"12","key":"2633_CR19","doi-asserted-by":"crossref","first-page":"1882","DOI":"10.1109\/LSP.2016.2618776","volume":"23","author":"L Yu","year":"2016","unstructured":"Yu, L., Xun, C., Ward, R.K., Wang, J.Z.: Image fusion with convolutional sparse representation. IEEE Signal Process. Lett. 23(12), 1882\u20131886 (2016)","journal-title":"IEEE Signal Process. Lett."},{"issue":"5","key":"2633_CR20","doi-asserted-by":"crossref","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Xiao-Jun, W.: Densefuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"2633_CR21","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Wei, Y., Liang, P., Li, C., Jiang, J.: Fusiongan: a generative adversarial network for infrared and visible image fusion. Inf. Fus. 48, 11\u201326 (2019)","journal-title":"Inf. Fus."},{"issue":"1","key":"2633_CR22","doi-asserted-by":"crossref","first-page":"75","DOI":"10.1016\/S1000-9361(08)60071-0","volume":"22","author":"L Kun","year":"2009","unstructured":"Kun, L., Lei, G., Huihui, L., Jingsong, C.: Fusion of infrared and visible light images based on region segmentation. Chin. J. Aeronaut. 22(1), 75\u201380 (2009)","journal-title":"Chin. J. Aeronaut."},{"key":"2633_CR23","doi-asserted-by":"crossref","first-page":"100","DOI":"10.1016\/j.inffus.2016.02.001","volume":"31","author":"J Ma","year":"2016","unstructured":"Ma, J., Chen, C., Li, C., Huang, J.: Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fus. 31, 100\u2013109 (2016)","journal-title":"Inf. Fus."},{"key":"2633_CR24","doi-asserted-by":"crossref","first-page":"218","DOI":"10.1016\/j.jvcir.2016.06.026","volume":"40","author":"J Adu","year":"2016","unstructured":"Adu, J., Xie, S., Gan, J.: Image fusion based on visual salient features and the cross-contrast. J. Vis. Commun. Image Represent. 40, 218\u2013224 (2016)","journal-title":"J. Vis. Commun. Image Represent."},{"issue":"12","key":"2633_CR25","doi-asserted-by":"crossref","first-page":"2706","DOI":"10.1109\/TMM.2017.2711422","volume":"19","author":"H Hai-Miao","year":"2017","unstructured":"Hai-Miao, H., Jiawei, W., Li, B., Guo, Q., Zheng, J.: An adaptive fusion algorithm for visible and infrared videos based on entropy and the cumulative distribution of gray levels. IEEE Trans. Multimedia 19(12), 2706\u20132719 (2017)","journal-title":"IEEE Trans. Multimedia"},{"issue":"4","key":"2633_CR26","doi-asserted-by":"crossref","first-page":"866","DOI":"10.1109\/TMM.2017.2760100","volume":"20","author":"W Zhao","year":"2018","unstructured":"Zhao, W., Huimin, L., Wang, D.: Multisensor image fusion and enhancement in spectral total variation domain. IEEE Trans. Multimedia 20(4), 866\u2013879 (2018)","journal-title":"IEEE Trans. Multimedia"},{"issue":"9","key":"2633_CR27","doi-asserted-by":"crossref","first-page":"2283","DOI":"10.1109\/TIM.2017.2700198","volume":"66","author":"W Zhao","year":"2017","unstructured":"Zhao, W., Huchuan, L.: Medical image fusion and denoising with alternating sequential filter and adaptive fractional order total variation. IEEE Trans. Instrum. Meas. 66(9), 2283\u20132294 (2017)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"2633_CR28","doi-asserted-by":"crossref","first-page":"1261","DOI":"10.1109\/TIP.2020.3043125","volume":"30","author":"R Liu","year":"2020","unstructured":"Liu, R., Liu, J., Jiang, Z., Fan, X., Luo, Z.: A bilevel integrated model with data-driven layer ensemble for multi-modality image fusion. IEEE Trans. Image Process. 30, 1261\u20131274 (2020)","journal-title":"IEEE Trans. Image Process."},{"issue":"4","key":"2633_CR29","doi-asserted-by":"crossref","first-page":"884","DOI":"10.1109\/TIM.2009.2026612","volume":"59","author":"B Yang","year":"2010","unstructured":"Yang, B., Li, S.: Multifocus image fusion and restoration with sparse representation. IEEE Trans. Instrum. Meas. 59(4), 884\u2013892 (2010)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"2633_CR30","doi-asserted-by":"crossref","first-page":"1818","DOI":"10.1109\/LSP.2021.3109818","volume":"28","author":"J Liu","year":"2021","unstructured":"Liu, J., Yuhui, W., Huang, Z., Liu, R., Fan, X.: SMoA: Searching a modality-oriented architecture for infrared and visible image fusion. IEEE Signal Process. Lett. 28, 1818\u20131822 (2021)","journal-title":"IEEE Signal Process. Lett."},{"key":"2633_CR31","doi-asserted-by":"crossref","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu, Y., Xun, C., Peng, H., Wang, Z.: Multi-focus image fusion with a deep convolutional neural network. Inf. Fus. 36, 191\u2013207 (2017)","journal-title":"Inf. Fus."},{"issue":"2","key":"2633_CR32","doi-asserted-by":"crossref","first-page":"143","DOI":"10.1016\/j.inffus.2006.02.001","volume":"8","author":"F Nencini","year":"2007","unstructured":"Nencini, F., Garzelli, A., Baronti, S., Alparone, L.: Remote sensing image fusion using the curvelet transform. Inf. Fus. 8(2), 143\u2013156 (2007)","journal-title":"Inf. Fus."},{"key":"2633_CR33","doi-asserted-by":"crossref","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","volume":"82","author":"J Ma","year":"2017","unstructured":"Ma, J., Zhou, Z., Wang, B., Zong, H.: Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 82, 8\u201317 (2017)","journal-title":"Infrared Phys. Technol."},{"key":"2633_CR34","doi-asserted-by":"crossref","first-page":"375","DOI":"10.1016\/j.compeleceng.2016.09.019","volume":"62","author":"F Meng","year":"2017","unstructured":"Meng, F., Song, M., Guo, B., Shi, R., Shan, D.: Image fusion based on object region detection and non-subsampled contourlet transform. Comput. Electr. Eng. 62, 375\u2013383 (2017)","journal-title":"Comput. Electr. Eng."},{"key":"2633_CR35","doi-asserted-by":"crossref","first-page":"282","DOI":"10.1016\/j.infrared.2017.01.013","volume":"81","author":"P Zhu","year":"2017","unstructured":"Zhu, P., Ma, X., Huang, Z.: Fusion of infrared-visible images using improved multi-scale top-hat transform and suitable fusion rules. Infrared Phys. Technol. 81, 282\u2013295 (2017)","journal-title":"Infrared Phys. Technol."},{"issue":"4","key":"2633_CR36","doi-asserted-by":"crossref","first-page":"1172","DOI":"10.1109\/TBME.2018.2869432","volume":"66","author":"W Li","year":"2018","unstructured":"Li, W., Jiao, D., Zhao, Z., Long, J.: Fusion of medical sensors using adaptive cloud model in local Laplacian pyramid domain. IEEE Trans. Biomed. Eng. 66(4), 1172\u20131183 (2018)","journal-title":"IEEE Trans. Biomed. Eng."},{"key":"2633_CR37","doi-asserted-by":"crossref","first-page":"6724","DOI":"10.1109\/ACCESS.2017.2685178","volume":"5","author":"P Chai","year":"2017","unstructured":"Chai, P., Luo, X., Zhang, Z.: Image fusion using quaternion wavelet transform and multiple features. IEEE Access 5, 6724\u20136734 (2017)","journal-title":"IEEE Access"},{"key":"2633_CR38","doi-asserted-by":"crossref","first-page":"286","DOI":"10.1016\/j.infrared.2015.10.004","volume":"73","author":"B Zhang","year":"2015","unstructured":"Zhang, B., Xiaoqi, L., Pei, H., Zhao, Y.: A fusion algorithm for infrared and visible images based on saliency analysis and non-subsampled shearlet transform. Infrared Phys. Technol. 73, 286\u2013297 (2015)","journal-title":"Infrared Phys. Technol."},{"issue":"5","key":"2633_CR39","doi-asserted-by":"crossref","first-page":"2898","DOI":"10.1109\/TIP.2012.2183140","volume":"21","author":"J Liang","year":"2012","unstructured":"Liang, J., He, Y., Liu, D., Zeng, X.: Image fusion using higher order singular value decomposition. IEEE Trans. Image Process. 21(5), 2898\u20132909 (2012)","journal-title":"IEEE Trans. Image Process."},{"key":"2633_CR40","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1016\/j.neucom.2014.07.003","volume":"148","author":"H Yin","year":"2015","unstructured":"Yin, H.: Sparse representation with learned multiscale dictionary for image fusion. Neurocomputing 148, 600\u2013610 (2015)","journal-title":"Neurocomputing"},{"key":"2633_CR41","doi-asserted-by":"crossref","first-page":"130","DOI":"10.1016\/j.patcog.2018.02.005","volume":"79","author":"H Li","year":"2018","unstructured":"Li, H., He, X., Tao, D., Tang, Y., Wang, R.: Joint medical image fusion, denoising and enhancement via discriminative low-rank sparse dictionaries learning. Pattern Recogn. 79, 130\u2013146 (2018)","journal-title":"Pattern Recogn."},{"issue":"12","key":"2633_CR42","doi-asserted-by":"crossref","first-page":"1882","DOI":"10.1109\/LSP.2016.2618776","volume":"23","author":"L Yu","year":"2016","unstructured":"Yu, L., Xun, C., Ward, R.K., Wang, J.Z.: Image fusion with convolutional sparse representation. IEEE Signal Process. Lett. 23(12), 1882\u20131886 (2016)","journal-title":"IEEE Signal Process. Lett."},{"issue":"1","key":"2633_CR43","doi-asserted-by":"crossref","first-page":"105","DOI":"10.1109\/TCSVT.2021.3056725","volume":"32","author":"J Liu","year":"2021","unstructured":"Liu, J., Fan, X., Jiang, J., Liu, R., Luo, Z.: Learning a deep multi-scale feature ensemble and an edge-attention guidance for image fusion. IEEE Trans. Circuits Syst. Video Technol. 32(1), 105\u2013119 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"2633_CR44","doi-asserted-by":"crossref","unstructured":"Zhong, J., Yang, B., Li, Y., Zhong, F., Chen, Z.: Image fusion and super-resolution with convolutional neural network. In: Chinese Conference on Pattern Recognition, pp. 78\u201388. Springer, Berlin (2016)","DOI":"10.1007\/978-981-10-3005-5_7"},{"issue":"03","key":"2633_CR45","doi-asserted-by":"crossref","first-page":"1850018","DOI":"10.1142\/S0219691318500182","volume":"16","author":"Y Liu","year":"2018","unstructured":"Liu, Y., Chen, X., Cheng, J., Peng, H., Wang, Z.: Infrared and visible image fusion with convolutional neural networks. Int. J. Wavelets Multiresolution Inf. Process. 16(03), 1850018 (2018)","journal-title":"Int. J. Wavelets Multiresolution Inf. Process."},{"issue":"3","key":"2633_CR46","doi-asserted-by":"crossref","first-page":"1041","DOI":"10.1016\/j.asoc.2011.11.020","volume":"12","author":"J Saeedi","year":"2012","unstructured":"Saeedi, J., Faez, K.: Infrared and visible image fusion using fuzzy logic and population-based optimization. Appl. Soft Comput. 12(3), 1041\u20131054 (2012)","journal-title":"Appl. Soft Comput."},{"issue":"11","key":"2633_CR47","doi-asserted-by":"crossref","first-page":"1961","DOI":"10.1364\/JOSAA.34.001961","volume":"34","author":"H Guo","year":"2017","unstructured":"Guo, H., Ma, Y., Mei, X., Ma, J.: Infrared and visible image fusion based on total variation and augmented Lagrangian. J. Opt. Soc. Am. A 34(11), 1961\u20131968 (2017)","journal-title":"J. Opt. Soc. Am. A"},{"issue":"1","key":"2633_CR48","doi-asserted-by":"crossref","DOI":"10.1117\/1.JEI.25.1.013016","volume":"25","author":"T Shibata","year":"2016","unstructured":"Shibata, T., Tanaka, M., Okutomi, M.: Versatile visible and near-infrared image fusion based on high visibility area selection. J. Electron. Imaging 25(1), 013016 (2016)","journal-title":"J. Electron. Imaging"},{"issue":"2","key":"2633_CR49","doi-asserted-by":"crossref","first-page":"127","DOI":"10.1016\/j.inffus.2011.08.002","volume":"14","author":"Y Han","year":"2013","unstructured":"Han, Y., Cai, Y., Cao, Y., Xiaoming, X.: A new image fusion performance metric based on visual information fidelity. Inf. Fus. 14(2), 127\u2013135 (2013)","journal-title":"Inf. Fus."},{"key":"2633_CR50","doi-asserted-by":"crossref","first-page":"546","DOI":"10.1016\/j.infrared.2016.04.015","volume":"76","author":"X Bai","year":"2016","unstructured":"Bai, X.: Morphological center operator based infrared and visible image fusion through correlation coefficient. Infrared Phys. Technol. 76, 546\u2013554 (2016)","journal-title":"Infrared Phys. Technol."},{"key":"2633_CR51","doi-asserted-by":"crossref","first-page":"174","DOI":"10.1016\/j.infrared.2016.02.005","volume":"76","author":"H Li","year":"2016","unstructured":"Li, H., Qiu, H., Zhengtao, Y., Zhang, Y.: Infrared and visible image fusion scheme based on NSCT and low-level visual features. Infrared Phys. Technol. 76, 174\u2013184 (2016)","journal-title":"Infrared Phys. Technol."},{"key":"2633_CR52","doi-asserted-by":"crossref","unstructured":"Wang, A., Jiang, J., Zhang, H.: Multi-sensor image decision level fusion detection algorithm based on DS evidence theory. In: 2014 Fourth International Conference on Instrumentation and Measurement, Computer, Communication and Control, pp. 620\u2013623. IEEE (2014)","DOI":"10.1109\/IMCCC.2014.132"},{"key":"2633_CR53","unstructured":"Lahoud, F., S\u00fcsstrunk, S.: Fast and efficient zero-learning image fusion. arXiv preprint arXiv:1905.03590 (2019)"},{"key":"2633_CR54","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: Advances in Neural Information Processing Systems, pp. 2672\u20132680 (2014)"},{"key":"2633_CR55","unstructured":"Nichol, A., Achiam, J., Schulman, J.: On first-order meta-learning algorithms. arXiv preprint arXiv:1803.02999 (2018)"},{"issue":"7","key":"2633_CR56","doi-asserted-by":"crossref","first-page":"3142","DOI":"10.1109\/TIP.2017.2662206","volume":"26","author":"K Zhang","year":"2016","unstructured":"Zhang, K., Zuo, W., Chen, Y., Meng, D., Zhang, L.: Beyond a gaussian denoiser: residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 26(7), 3142\u20133155 (2016)","journal-title":"IEEE Trans. Image Process."},{"issue":"7","key":"2633_CR57","doi-asserted-by":"crossref","first-page":"3142","DOI":"10.1109\/TIP.2017.2662206","volume":"26","author":"K Zhang","year":"2017","unstructured":"Zhang, K., Zuo, W., Chen, Y., Meng, D., Zhang, L.: Beyond a gaussian denoiser: Residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 26(7), 3142\u20133155 (2017)","journal-title":"IEEE Trans. Image Process."},{"issue":"3","key":"2633_CR58","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., Deng, J., Hao, S., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al.: Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 115(3), 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vis."},{"issue":"12","key":"2633_CR59","doi-asserted-by":"crossref","first-page":"1890","DOI":"10.1016\/j.aeue.2015.09.004","volume":"69","author":"V Aslantas","year":"2015","unstructured":"Aslantas, V., Bendes, E.: A new image quality metric for image fusion: the sum of the correlations of differences. AEU-Int. J. Electron. Commun. 69(12), 1890\u20131896 (2015)","journal-title":"AEU-Int. J. Electron. Commun."},{"issue":"4","key":"2633_CR60","doi-asserted-by":"crossref","first-page":"308","DOI":"10.1049\/el:20000267","volume":"36","author":"CS Xydeas","year":"2000","unstructured":"Xydeas, C.S., Vladimir, P.: Objective image fusion performance measure. Electron. Lett. 36(4), 308\u2013309 (2000)","journal-title":"Electron. Lett."},{"key":"2633_CR61","doi-asserted-by":"crossref","first-page":"94","DOI":"10.1016\/j.infrared.2017.04.018","volume":"83","author":"CH Liu","year":"2017","unstructured":"Liu, C.H., Qi, Y., Ding, W.R.: Infrared and visible image fusion method based on saliency detection in sparse domain. Infrared Phys. Technol. 83, 94\u2013102 (2017)","journal-title":"Infrared Phys. Technol."},{"key":"2633_CR62","doi-asserted-by":"crossref","unstructured":"Qinglei, D., Han, X., Ma, Y., Huang, J., Fan, F.: Fusing infrared and visible images of different resolutions via total variation model. Sensors 18(11), 3827 (2018)","DOI":"10.3390\/s18113827"},{"issue":"3","key":"2633_CR63","doi-asserted-by":"crossref","first-page":"410","DOI":"10.1006\/nimg.2000.0707","volume":"13","author":"V Barra","year":"2001","unstructured":"Barra, V., Boire, J.-Y.: A general framework for the fusion of anatomical and functional medical images. Neuroimage 13(3), 410\u2013424 (2001)","journal-title":"Neuroimage"},{"issue":"5","key":"2633_CR64","doi-asserted-by":"crossref","first-page":"347","DOI":"10.1049\/iet-ipr.2014.0311","volume":"9","author":"Y Liu","year":"2014","unstructured":"Liu, Y., Wang, Z.: Simultaneous image fusion and denoising with adaptive sparse representation. Image Process. IET 9(5), 347\u2013357 (2014)","journal-title":"Image Process. IET"},{"issue":"1","key":"2633_CR65","doi-asserted-by":"crossref","first-page":"49","DOI":"10.1109\/TIM.2018.2838778","volume":"68","author":"M Yin","year":"2019","unstructured":"Yin, M., Liu, X., Liu, Y., Chen, X.: Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Instrum. Meas. Soc. 68(1), 49\u201364 (2019)","journal-title":"IEEE Instrum. Meas. Soc."},{"key":"2633_CR66","doi-asserted-by":"crossref","unstructured":"Brown, M., S\u00fcsstrunk, S.: Multi-spectral sift for scene category recognition. In: CVPR, pp. 177\u2013184 (2011)","DOI":"10.1109\/CVPR.2011.5995637"},{"key":"2633_CR67","unstructured":"Tomasi, C., Manduchi, R.: Bilateral filtering for gray and color images. In: Sixth International Conference on Computer Vision (IEEE Cat. No. 98CH36271), pp. 839\u2013846. IEEE (1998)"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-022-02633-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-022-02633-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-022-02633-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,2]],"date-time":"2024-10-02T15:42:24Z","timestamp":1727883744000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-022-02633-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,8,25]]},"references-count":67,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2023,10]]}},"alternative-id":["2633"],"URL":"https:\/\/doi.org\/10.1007\/s00371-022-02633-9","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2022,8,25]]},"assertion":[{"value":"15 April 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 August 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We declare that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work, there is no professional or other personal interest of any nature or kind in any product, service, and\/or company that could be construed as influencing the position presented in, or the review of, the manuscript entitled, \u201cA Unified Image Fusion Framework with Flexible Bilevel Paradigm Integration.\u201d","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}