{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T04:03:40Z","timestamp":1769745820539,"version":"3.49.0"},"reference-count":39,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2022,4,29]],"date-time":"2022-04-29T00:00:00Z","timestamp":1651190400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,4,29]],"date-time":"2022-04-29T00:00:00Z","timestamp":1651190400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2023,1]]},"DOI":"10.1007\/s10489-022-03658-2","type":"journal-article","created":{"date-parts":[[2022,4,29]],"date-time":"2022-04-29T15:03:58Z","timestamp":1651244638000},"page":"1452-1468","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["A new multi-focus image fusion method based on multi-classification focus learning and multi-scale decomposition"],"prefix":"10.1007","volume":"53","author":[{"given":"Lifeng","family":"Ma","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6659-2844","authenticated-orcid":false,"given":"Yanxiang","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Jiaqi","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zhijie","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Wenhao","family":"Sun","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,29]]},"reference":[{"key":"3658_CR1","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.05.004","volume":"33","author":"S Li","year":"2017","unstructured":"Li S, Kang X, Fang L, Hu J, Yin H (2017) Pixel-level image fusion: a survey of the state of the art. Information Fusion 33:100\u2013112","journal-title":"Information Fusion"},{"key":"3658_CR2","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1016\/j.inffus.2020.06.013","volume":"64","author":"Y Liu","year":"2020","unstructured":"Liu Y, Wang L, Cheng J, Li C, Chen X (2020) Multi-focus image fusion: a survey of the state of the art. Information Fusion 64:71\u201391","journal-title":"Information Fusion"},{"key":"3658_CR3","doi-asserted-by":"publisher","first-page":"5735","DOI":"10.1007\/s10462-021-09961-7","volume":"54","author":"S Bhat","year":"2021","unstructured":"Bhat S, Koundal D (2021) Multi-focus image fusion techniques: a survey. Artif Intell Rev 54:5735\u20135787","journal-title":"Artif Intell Rev"},{"key":"3658_CR4","first-page":"2021","volume":"1-1","author":"X Zhang","year":"2021","unstructured":"Zhang X (2021) Deep learning-based multi-focus image fusion: a survey and a comparative study. IEEE Trans Pattern Anal Mach Intell 1-1:2021","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"3658_CR5","doi-asserted-by":"crossref","unstructured":"Zhang H, Xu H, Tian X et al. (2021) Image fusion meets deep learning: a survey and perspective, information fusion, 76: 323\u2013336","DOI":"10.1016\/j.inffus.2021.06.008"},{"key":"3658_CR6","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1016\/j.inffus.2017.10.007","volume":"42","author":"Y Liu","year":"2018","unstructured":"Liu Y, Chen X, Wang Z, Wang ZJ, Ward RK, Wang X (2018) Deep learning for pixel-level image fusion: recent advances and future prospects. Information Fusion 42:158\u2013173","journal-title":"Information Fusion"},{"key":"3658_CR7","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu Y, Chen X, Peng H, Wang Z (2017) Multi-focus image fusion with a deep convolutional neural network. Information Fusion 36:191\u2013207","journal-title":"Information Fusion"},{"key":"3658_CR8","doi-asserted-by":"publisher","first-page":"8668","DOI":"10.1109\/TIP.2020.3018261","volume":"29","author":"H Ma","year":"2020","unstructured":"Ma H, Liao Q, Zhang J, Liu S, Xue JH (2020) An \u03b1-matte boundary defocus model-based cascaded network for multi-focus image fusion. IEEE Transaction on Image Processing 29:8668\u20138679","journal-title":"IEEE Transaction on Image Processing"},{"key":"3658_CR9","doi-asserted-by":"publisher","first-page":"201","DOI":"10.1016\/j.inffus.2019.02.003","volume":"51","author":"M Amin-Naji","year":"2019","unstructured":"Amin-Naji M, Aghagolzadeh A, Ezoji M (2019) Ensemble of CNN for multi-focus image fusion. Information Fusion 51:201\u2013214","journal-title":"Information Fusion"},{"key":"3658_CR10","doi-asserted-by":"publisher","first-page":"125","DOI":"10.1016\/j.ins.2017.12.043","volume":"433-434","author":"H Tang","year":"2018","unstructured":"Tang H, Xiao B, Li W, Wang G (2018) Pixel convolutional neural network for multi-focus image fusion. Inf Sci 433-434:125\u2013141","journal-title":"Inf Sci"},{"key":"3658_CR11","doi-asserted-by":"crossref","unstructured":"Guo X, Nie R. Cao J, et al. (2018) Fully convolutional network based multifocus image fusion, neural computation, 30(7):1\u201326","DOI":"10.1162\/neco_a_01098"},{"key":"3658_CR12","doi-asserted-by":"publisher","first-page":"4816","DOI":"10.1109\/TIP.2020.2976190","volume":"29","author":"J Li","year":"2020","unstructured":"Li J, Guo X, Lu G (2020) DRPL: deep regression pair learning for multi-focus image fusion. IEEE Transaction on Image Processing 29:4816\u20134831","journal-title":"IEEE Transaction on Image Processing"},{"key":"3658_CR13","doi-asserted-by":"publisher","first-page":"1561","DOI":"10.1109\/TCI.2020.3039564","volume":"6","author":"S Xu","year":"2020","unstructured":"Xu S, Ji L, Wang Z, Li P, Sun K, Zhang C, Zhang J (2020) Towards reducing severe defocus spread effects for multi-focus image fusion via an optimization based strategy. IEEE Transactions on Computational Imaging 6:1561\u20131570","journal-title":"IEEE Transactions on Computational Imaging"},{"key":"3658_CR14","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1016\/j.inffus.2014.09.004","volume":"24","author":"Y Liu","year":"2015","unstructured":"Liu Y, Liu S, Wang Z (2015) A general framework for image fusion based on multi-scale transform and sparse representation. Information Fusion 24:147\u2013164","journal-title":"Information Fusion"},{"key":"3658_CR15","doi-asserted-by":"publisher","first-page":"116572","DOI":"10.1016\/j.image.2021.116572","volume":"101","author":"Y Hu","year":"2022","unstructured":"Hu Y, Zhang B, Zhang Y, Chen Z (2022) A novel multi-focus image fusion method based on joint regularization optimization layering and sparse representation. Signal Process Image Commun 101:116572","journal-title":"Signal Process Image Commun"},{"key":"3658_CR16","doi-asserted-by":"publisher","first-page":"57","DOI":"10.1016\/j.inffus.2017.05.006","volume":"40","author":"Q Zhang","year":"2018","unstructured":"Zhang Q, Liu Y, Blum R et al (2018) Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: a review. Information Fusion 40:57\u201375","journal-title":"Information Fusion"},{"key":"3658_CR17","doi-asserted-by":"publisher","first-page":"370","DOI":"10.1016\/j.patrec.2020.08.002","volume":"138","author":"J Zhang","year":"2020","unstructured":"Zhang J, Liao Q, Liu S, Ma H, Yang W, Xue JH (2020) Real-MFF: a large realistic multi-focus image dataset with ground truth. Pattern Recogn Lett 138:370\u2013377","journal-title":"Pattern Recogn Lett"},{"issue":"04","key":"3658_CR18","first-page":"1","volume":"52","author":"L Jiang","year":"2022","unstructured":"Jiang L, Fan H, Li J (2022) A multi-focus image fusion method based on attention mechanism and supervised learning, applied intelligence. Vol. 52(04):1\u201319","journal-title":"Vol."},{"key":"3658_CR19","volume-title":"J. Ma and Z. Le et al","author":"H Xu","year":"2020","unstructured":"Xu H (2020) FusionDN: A Unified Densely Connected Network for Image Fusion. In: J. Ma and Z. Le et al. FusionDN, A Unified Densely Connected Network for Image Fusion, Proceedings of the AAAI Conference on Artificial Intelligence"},{"key":"3658_CR20","doi-asserted-by":"publisher","first-page":"26","DOI":"10.1016\/j.imavis.2019.03.001","volume":"85","author":"H Mustafa","year":"2019","unstructured":"Mustafa H, Yang J, Zareapoor M (2019) Multi-scale convolutional neural network for multi-focus image fusion. Image Vis Comput 85:26\u201335","journal-title":"Image Vis Comput"},{"key":"3658_CR21","doi-asserted-by":"publisher","first-page":"115864","DOI":"10.1016\/j.image.2020.115864","volume":"85","author":"H Mustafa","year":"2020","unstructured":"Mustafa H, Zareapoor M, Yang J (2020) MLDNet: multi-level dense network for multi-focus image fusion. Signal Process Image Commun 85:115864","journal-title":"Signal Process Image Commun"},{"key":"3658_CR22","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"H Zhang","year":"2021","unstructured":"Zhang H, Le Z, Shao Z et al (2021) MFF-GAN: an unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Information Fusion 66:40\u201353","journal-title":"Information Fusion"},{"key":"3658_CR23","first-page":"2019","volume":"21","author":"X Guo","year":"1982","unstructured":"Guo X, Nie R, Cao J et al (1982-1996) FuseGAN: learning to fuse multi-focus image via conditional generative adversarial network, IEEE transactions on multimedia. Vol. 21:2019","journal-title":"Vol."},{"key":"3658_CR24","doi-asserted-by":"publisher","first-page":"135284","DOI":"10.1109\/ACCESS.2020.3010542","volume":"8","author":"Z Duan","year":"2020","unstructured":"Duan Z, Zhang T, Tan J, Luo X (2020) Non-local multi-focus image fusion with recurrent neural networks. IEEE Access 8:135284\u2013135295","journal-title":"IEEE Access"},{"key":"3658_CR25","doi-asserted-by":"publisher","first-page":"114385","DOI":"10.1109\/ACCESS.2019.2935006","volume":"7","author":"R Lai","year":"2019","unstructured":"Lai R, Li Y, Guan J, Xiong A (2019) Multi-scale visual attention deep convolutional neural network for multi-focus image fusion. IEEE Access 7:114385\u2013114399","journal-title":"IEEE Access"},{"key":"3658_CR26","doi-asserted-by":"crossref","unstructured":"Prakash A, Chitta K, Geiger A (2021) Multi-modal fusion transformer for end-to-end autonomous driving, CVPR","DOI":"10.1109\/CVPR46437.2021.00700"},{"key":"3658_CR27","doi-asserted-by":"publisher","first-page":"16040","DOI":"10.1109\/ACCESS.2017.2735865","volume":"5","author":"A Dogra","year":"2017","unstructured":"Dogra A, Goyal B, Agrawal S (2017) From multi-scale decomposition to non-multi-scale decomposition methods: a comprehensive survey of image fusion techniques and its applications. IEEE access 5:16040\u201316067","journal-title":"IEEE access"},{"issue":"99","key":"3658_CR28","doi-asserted-by":"publisher","first-page":"103328","DOI":"10.1016\/j.jvcir.2021.103328","volume":"81","author":"X Ma","year":"2021","unstructured":"Ma X, Wang Z, Hu S (2021) Multi-focus image fusion based on multi-scale sparse representation. Journal of Visual Communication and Image Representation 81(99):103328","journal-title":"Journal of Visual Communication and Image Representation"},{"key":"3658_CR29","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition, ImageNet large scale visual recognition challenge (ILSVRC) workshop"},{"key":"3658_CR30","doi-asserted-by":"publisher","first-page":"493","DOI":"10.1016\/j.patrec.2006.09.005","volume":"28","author":"W Huang","year":"2007","unstructured":"Huang W, Jing Z (2007) Evaluation of focus measures in multi-focus image fusion. Pattern Recogn 28:493\u2013500","journal-title":"Pattern Recogn"},{"key":"3658_CR31","first-page":"2016","volume":"23","author":"Y Liu","year":"1882","unstructured":"Liu Y, Chen X, Ward R et al (1882-1886) Image fusion with convolutional sparse representation. IEEE Signal Processing Letters 23:2016\u20131886","journal-title":"IEEE Signal Processing Letters"},{"key":"3658_CR32","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang Y, Liu Y, Sun P, Yan H, Zhao X, Zhang L (2020) IFCNN: a general image fusion framework based on convolutional neural network. Information Fusion 54:99\u2013118","journal-title":"Information Fusion"},{"issue":"1","key":"3658_CR33","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu H, Ma J, Jiang J et al (2020) U2Fusion: a unified unsupervised image fusion network. IEEE transactions on pattern analysis and machine intelligence. 44(1):502\u2013518","journal-title":"IEEE transactions on pattern analysis and machine intelligence."},{"key":"3658_CR34","doi-asserted-by":"publisher","first-page":"313","DOI":"10.1049\/el:20020212","volume":"38","author":"G Qu","year":"2002","unstructured":"Qu G, Zhang D (2002) Information measure for performance of image fusion. Electron Lett 38:313\u2013315","journal-title":"Electron Lett"},{"key":"3658_CR35","doi-asserted-by":"publisher","first-page":"308","DOI":"10.1049\/el:20000267","volume":"36","author":"C Xydeas","year":"2000","unstructured":"Xydeas C, Petrovic V (2000) Objective image fusion performance measure. Electron Lett 36:308\u2013309","journal-title":"Electron Lett"},{"key":"3658_CR36","first-page":"1433","volume":"3","author":"J Zhao","year":"2007","unstructured":"Zhao J, Laganiere R, Liu Z (2007) Performance assessment of combinative pixel-level image fusion based on an absolute feature measurement. International Journal of Innovative Computing, Information and Control 3:1433\u20131447","journal-title":"International Journal of Innovative Computing, Information and Control"},{"key":"3658_CR37","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1016\/j.inffus.2006.09.001","volume":"9","author":"C Yang","year":"2008","unstructured":"Yang C, Zhang J, Wang X et al (2008) A novel similarity based quality metric for image fusion. Information Fusion 9:156\u2013160","journal-title":"Information Fusion"},{"key":"3658_CR38","doi-asserted-by":"publisher","first-page":"1421","DOI":"10.1016\/j.imavis.2007.12.002","volume":"27","author":"Y Chen","year":"2009","unstructured":"Chen Y, Blum R (2009) A new automated quality assessment algorithm for image fusion. Image and Vision Computing 27:1421\u20131432","journal-title":"Image and Vision Computing"},{"key":"3658_CR39","doi-asserted-by":"publisher","first-page":"1032","DOI":"10.1049\/ipr2.12083","volume":"15","author":"Y Hu","year":"2021","unstructured":"Hu Y, Zhang B, Zhang J, Gao Q (2021) Multi-focus image fusion evaluation based on jointly sparse representation and atom focus measure. IET Image Process 15:1032\u20131041","journal-title":"IET Image Process"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03658-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-022-03658-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03658-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,4]],"date-time":"2023-01-04T04:41:56Z","timestamp":1672807316000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-022-03658-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,29]]},"references-count":39,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2023,1]]}},"alternative-id":["3658"],"URL":"https:\/\/doi.org\/10.1007\/s10489-022-03658-2","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,29]]},"assertion":[{"value":"18 April 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 April 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}