{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,21]],"date-time":"2025-05-21T12:30:33Z","timestamp":1747830633167,"version":"3.33.0"},"reference-count":40,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,4,12]],"date-time":"2024-04-12T00:00:00Z","timestamp":1712880000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,4,12]],"date-time":"2024-04-12T00:00:00Z","timestamp":1712880000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-024-19088-w","type":"journal-article","created":{"date-parts":[[2024,4,12]],"date-time":"2024-04-12T05:01:31Z","timestamp":1712898091000},"page":"1239-1259","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Mutli-focus image fusion based on guided filter and image matting network"],"prefix":"10.1007","volume":"84","author":[{"given":"Puchao","family":"Zhu","sequence":"first","affiliation":[]},{"given":"Xue","family":"Li","sequence":"additional","affiliation":[]},{"given":"Puming","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Shaowen","family":"Yao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,4,12]]},"reference":[{"key":"19088_CR1","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1016\/j.compeleceng.2016.03.011","volume":"51","author":"M Abdipour","year":"2016","unstructured":"Abdipour M, Nooshyar M (2016) Multi-focus image fusion using sharpness criteria for visual sensor networks in wavelet domain. Comput Electr Eng 51:74\u201388","journal-title":"Comput Electr Eng"},{"key":"19088_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2021.3090410","volume":"60","author":"UA Bhatti","year":"2021","unstructured":"Bhatti UA, Yu Z, Chanussot J, Zeeshan Z, Yuan L, Luo W, Nawaz SA, Bhatti MA, Ain QU, Mehmood A (2021) Local similarity-based spatial-spectral fusion hyperspectral image classification with deep cnn and gabor filtering. IEEE Trans Geosci Remote Sens 60:1\u201315","journal-title":"IEEE Trans Geosci Remote Sens"},{"issue":"10","key":"19088_CR3","doi-asserted-by":"publisher","first-page":"1421","DOI":"10.1016\/j.imavis.2007.12.002","volume":"27","author":"Y Chen","year":"2009","unstructured":"Chen Y, Blum RS (2009) A new automated quality assessment algorithm for image fusion. Image Vis Comput 27(10):1421\u20131432","journal-title":"Image Vis Comput"},{"key":"19088_CR4","doi-asserted-by":"publisher","first-page":"199","DOI":"10.1016\/j.optcom.2014.12.032","volume":"341","author":"G Cui","year":"2015","unstructured":"Cui G, Feng H, Xu Z, Li Q, Chen Y (2015) Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition. Optics Commun 341:199\u2013209","journal-title":"Optics Commun"},{"issue":"3","key":"19088_CR5","doi-asserted-by":"publisher","first-page":"298","DOI":"10.1109\/TCI.2018.2849326","volume":"4","author":"H Haim","year":"2018","unstructured":"Haim H, Elmalem S, Giryes R, Bronstein AM, Marom E (2018) Depth estimation from a single image using deep learned phase coded mask. IEEE Transactions on Computational Imaging 4(3):298\u2013310","journal-title":"IEEE Transactions on Computational Imaging"},{"key":"19088_CR6","doi-asserted-by":"crossref","unstructured":"Hamza AB, Krim H (2001) A variational approach to maximum a posteriori estimation for image denoising. In: International Workshop on Energy Minimization Methods in Computer Vision and Pattern Recognition, pages 19\u201334. Springer","DOI":"10.1007\/3-540-44745-8_2"},{"issue":"6","key":"19088_CR7","doi-asserted-by":"publisher","first-page":"1397","DOI":"10.1109\/TPAMI.2012.213","volume":"35","author":"K He","year":"2012","unstructured":"He K, Sun J, Tang X (2012) Guided image filtering. IEEE Trans Pattern Anal Mach Intell 35(6):1397\u20131409","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"19088_CR8","doi-asserted-by":"publisher","first-page":"4685","DOI":"10.1007\/s00500-018-3118-9","volume":"23","author":"K He","year":"2019","unstructured":"He K, Zhou D, Zhang X, Nie R, Jin X (2019) Multi-focus image fusion combining focus-region-level partition and pulse-coupled neural network. Soft Comput 23:4685\u20134699","journal-title":"Soft Comput"},{"key":"19088_CR9","doi-asserted-by":"publisher","first-page":"123231","DOI":"10.1109\/ACCESS.2019.2933646","volume":"7","author":"Z Ji","year":"2019","unstructured":"Ji Z, Kang X, Zhang K, Duan P, Hao Q (2019) A two-stage multi-focus image fusion framework robust to image mis-registration. IEEE Access 7:123231\u2013123243","journal-title":"IEEE Access"},{"key":"19088_CR10","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.05.004","volume":"33","author":"S Li","year":"2017","unstructured":"Li S, Kang X, Fang L, Hu J, Yin H (2017) Pixel-level image fusion: A survey of the state of the art. Information Fusion 33:100\u2013112","journal-title":"Information Fusion"},{"issue":"5","key":"19088_CR11","doi-asserted-by":"publisher","first-page":"2898","DOI":"10.1109\/TIP.2012.2183140","volume":"21","author":"J Liang","year":"2012","unstructured":"Liang J, He Y, Liu D, Zeng X (2012) Image fusion using higher order singular value decomposition. IEEE Trans Image Process 21(5):2898\u20132909","journal-title":"IEEE Trans Image Process"},{"key":"19088_CR12","doi-asserted-by":"crossref","unstructured":"Lin T-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D (2014) Piotrdoll\u00e1r, andclawrencezitnick. microsoftcoco: Commonobjectsin context. In: European conference on computer vision, pages 740\u2013755","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"19088_CR13","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1016\/j.sigpro.2013.10.010","volume":"97","author":"Y Liu","year":"2014","unstructured":"Liu Y, Jin J, Wang Q, Shen Y, Dong X (2014) Region level based multi-focus image fusion using quaternion wavelet and normalized cut. Signal Processing 97:9\u201330","journal-title":"Signal Processing"},{"key":"19088_CR14","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu Y, Chen X, Peng H, Wang Z (2017) Multi-focus image fusion with a deep convolutional neural network. Information Fusion 36:191\u2013207","journal-title":"Information Fusion"},{"issue":"11","key":"19088_CR15","doi-asserted-by":"publisher","first-page":"5793","DOI":"10.1007\/s00521-020-05358-9","volume":"33","author":"B Ma","year":"2021","unstructured":"Ma B, Zhu Y, Yin X, Ban X, Huang H, Mukeshimana M (2021) Sesf-fuse: An unsupervised deep model for multi-focus image fusion. Neural Computing and Applications 33(11):5793\u20135804","journal-title":"Neural Computing and Applications"},{"issue":"7","key":"19088_CR16","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma J, Tang L, Fan F, Huang J, Mei X, Ma Y (2022) Swinfusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA Journal of Automatica Sinica 9(7):1200\u20131217","journal-title":"IEEE\/CAA Journal of Automatica Sinica"},{"key":"19088_CR17","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2014.10.004","volume":"25","author":"M Nejati","year":"2015","unstructured":"Nejati M, Samavi S, Shirani S (2015) Multi-focus image fusion using dictionary-based sparse representation. Information Fusion 25:72\u201384","journal-title":"Information Fusion"},{"issue":"12","key":"19088_CR18","doi-asserted-by":"publisher","first-page":"3674","DOI":"10.1109\/TGRS.2006.881758","volume":"44","author":"PS Pradhan","year":"2006","unstructured":"Pradhan PS, King RL, Younan NH, Holcomb DW (2006) Estimation of the number of decomposition levels for a wavelet-based multiresolution multisensor image fusion. IEEE Trans Geosci Remote Sens 44(12):3674\u20133686","journal-title":"IEEE Trans Geosci Remote Sens"},{"issue":"7","key":"19088_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1049\/el:20020212","volume":"38","author":"G Qu","year":"2002","unstructured":"Qu G, Zhang D, Yan P (2002) Information measure for performance of image fusion. Electron Lett 38(7):1","journal-title":"Electron Lett"},{"key":"19088_CR20","unstructured":"Rajalingam B,\u00a0Priya R (2018) Hybrid multimodality medical image fusion technique for feature enhancement in medical diagnosis. Int J Eng Sci Invent 2(Special issue):52\u201360"},{"key":"19088_CR21","doi-asserted-by":"publisher","DOI":"10.7717\/peerj-cs.654","volume":"7","author":"PN Srinivasu","year":"2021","unstructured":"Srinivasu PN, Balas VE (2021) Self-learning network-based segmentation for real-time brain mr images through haris. PeerJ Comput Sci 7:e654","journal-title":"PeerJ Comput Sci"},{"key":"19088_CR22","doi-asserted-by":"publisher","first-page":"125","DOI":"10.1016\/j.ins.2017.12.043","volume":"433","author":"H Tang","year":"2018","unstructured":"Tang H, Xiao B, Li W, Wang G (2018) Pixel convolutional neural network for multi-focus image fusion. Inf Sci 433:125\u2013141","journal-title":"Inf Sci"},{"issue":"12","key":"19088_CR23","doi-asserted-by":"publisher","first-page":"3413","DOI":"10.1109\/TMI.2021.3085712","volume":"40","author":"R Verma","year":"2021","unstructured":"Verma R, Kumar N, Patil A, Kurian NC, Rane S, Graham S, Vu QD, Zwager M, Raza SEA, Rajpoot N et al (2021) Monusac 2020: A multi-organ nuclei segmentation and classification challenge. IEEE Trans Med Imaging 40(12):3413\u20133423","journal-title":"IEEE Trans Med Imaging"},{"key":"19088_CR24","doi-asserted-by":"crossref","unstructured":"Wang J, Cohen MF et\u00a0al (2008) Image and video matting: a survey. Foundations and Trends\u00ae in Computer Graphics and Vision 3(2):97\u2013175","DOI":"10.1561\/0600000019"},{"issue":"3\u20134","key":"19088_CR25","first-page":"287","volume":"200","author":"Q Wang","year":"2005","unstructured":"Wang Q, Shen Y, Zhang JQ (2005) A nonlinear correlation measure for multivariable data set. Physica D: Nonlinear Phenomena 200(3\u20134):287\u2013295","journal-title":"Physica D: Nonlinear Phenomena"},{"key":"19088_CR26","unstructured":"Wang Z, Simoncelli EP, Bovik AC (2003) Multiscale structural similarity for image quality assessment. In: The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003, volume\u00a02, pages 1398\u20131402. Ieee"},{"issue":"1","key":"19088_CR27","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu H, Ma J, Jiang J, Guo X, Ling H (2020) U2fusion: A unified unsupervised image fusion network. IEEE Trans Pattern Anal Mach Intell 44(1):502\u2013518","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"19088_CR28","doi-asserted-by":"publisher","first-page":"12484","DOI":"10.1609\/aaai.v34i07.6936","volume":"34","author":"H Xu","year":"2020","unstructured":"Xu H, Ma J, Le Z, Jiang J, Guo X (2020) Fusiondn: A unified densely connected network for image fusion. Proceedings of the AAAI Conference on Artificial Intelligence 34:12484\u201312491","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"key":"19088_CR29","doi-asserted-by":"crossref","unstructured":"Xu N, Price B, Cohen S, Huang T (2017) Deep image matting. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2970\u20132979","DOI":"10.1109\/CVPR.2017.41"},{"key":"19088_CR30","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2020.102821","volume":"106","author":"Y Xu","year":"2020","unstructured":"Xu Y, Sun B, Yan X, Hu J, Chen M (2020) Multi-focus image fusion using learning based matting with sum of the gaussian-based modified laplacian. Digital Signal Process 106:102821","journal-title":"Digital Signal Process"},{"issue":"4","key":"19088_CR31","doi-asserted-by":"publisher","first-page":"308","DOI":"10.1049\/el:20000267","volume":"36","author":"CS Xydeas","year":"2000","unstructured":"Xydeas CS, Petrovic V et al (2000) Objective image fusion performance measure. Electron Lett 36(4):308\u2013309","journal-title":"Electron Lett"},{"issue":"5","key":"19088_CR32","first-page":"2824","volume":"15","author":"Y Yang","year":"2014","unstructured":"Yang Y, Tong S, Huang S, Lin P (2014) Multifocus image fusion based on nsct and focused area detection. IEEE Sensors J 15(5):2824\u20132838","journal-title":"IEEE Sensors J"},{"key":"19088_CR33","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"H Zhang","year":"2021","unstructured":"Zhang H, Le Z, Shao Z, Xu H, Ma J (2021) Mff-gan: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf Fusion 66:40\u201353","journal-title":"Inf Fusion"},{"key":"19088_CR34","doi-asserted-by":"publisher","first-page":"2761","DOI":"10.1007\/s11263-021-01501-8","volume":"129","author":"H Zhang","year":"2021","unstructured":"Zhang H, Ma J (2021) Sdnet: A versatile squeeze-and-decomposition network for real-time image fusion. Int J Comput Vis 129:2761\u20132785","journal-title":"Int J Comput Vis"},{"key":"19088_CR35","doi-asserted-by":"publisher","first-page":"12797","DOI":"10.1609\/aaai.v34i07.6975","volume":"34","author":"H Zhang","year":"2020","unstructured":"Zhang H, Xu H, Xiao Y, Guo X, Ma J (2020) Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. Proceedings of the AAAI Conference on Artificial Intelligence 34:12797\u201312804","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"issue":"7","key":"19088_CR36","doi-asserted-by":"publisher","first-page":"1334","DOI":"10.1016\/j.sigpro.2009.01.012","volume":"89","author":"Q Zhang","year":"2009","unstructured":"Zhang Q, Guo B-l (2009) Multifocus image fusion using the nonsubsampled contourlet transform. Signal Process 89(7):1334\u20131346","journal-title":"Signal Process"},{"key":"19088_CR37","doi-asserted-by":"publisher","first-page":"57","DOI":"10.1016\/j.inffus.2017.05.006","volume":"40","author":"Q Zhang","year":"2018","unstructured":"Zhang Q, Liu Y, Blum RS, Han J, Tao D (2018) Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: A review. Information Fusion 40:57\u201375","journal-title":"Information Fusion"},{"key":"19088_CR38","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang Y, Liu Y, Sun P, Yan H, Zhao X, Zhang L (2020) Ifcnn: A general image fusion framework based on convolutional neural network. Inf Fusion 54:99\u2013118","journal-title":"Inf Fusion"},{"issue":"6","key":"19088_CR39","first-page":"1433","volume":"3","author":"J Zhao","year":"2007","unstructured":"Zhao J, Laganiere R, Liu Z (2007) Performance assessment of combinative pixel-level image fusion based on an absolute feature measurement. Int J Innov Comput Inf Control 3(6):1433\u20131447","journal-title":"Int J Innov Comput Inf Control"},{"key":"19088_CR40","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2021.104699","volume":"136","author":"H Zunair","year":"2021","unstructured":"Zunair H, Hamza AB (2021) Sharp u-net: Depthwise convolutional network for biomedical image segmentation. Comput Biol Med 136:104699","journal-title":"Comput Biol Med"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19088-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-19088-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19088-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,28]],"date-time":"2025-01-28T13:07:40Z","timestamp":1738069660000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-19088-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,12]]},"references-count":40,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2025,1]]}},"alternative-id":["19088"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-19088-w","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,4,12]]},"assertion":[{"value":"5 September 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 February 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 March 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 April 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}]}}