{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T11:10:02Z","timestamp":1747998602754,"version":"3.41.0"},"reference-count":53,"publisher":"Springer Science and Business Media LLC","issue":"15","license":[{"start":{"date-parts":[[2024,6,15]],"date-time":"2024-06-15T00:00:00Z","timestamp":1718409600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,6,15]],"date-time":"2024-06-15T00:00:00Z","timestamp":1718409600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/100016229","name":"Basic Research Laboratory","doi-asserted-by":"publisher","award":["XXX-XXX-103-11-07"],"award-info":[{"award-number":["XXX-XXX-103-11-07"]}],"id":[{"id":"10.13039\/100016229","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-024-19470-8","type":"journal-article","created":{"date-parts":[[2024,6,15]],"date-time":"2024-06-15T06:02:12Z","timestamp":1718431332000},"page":"14407-14431","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["PTIFNet: Pseudo-Twin network for multi-focus image fusion"],"prefix":"10.1007","volume":"84","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1490-8733","authenticated-orcid":false,"given":"Pan","family":"Wu","sequence":"first","affiliation":[]},{"given":"Jin","family":"Tang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,15]]},"reference":[{"issue":"9","key":"19470_CR1","doi-asserted-by":"publisher","first-page":"1123","DOI":"10.1016\/j.patrec.2007.01.013","volume":"28","author":"H Wei","year":"2007","unstructured":"Wei H, Jing Z (2007) Multi-focus image fusion using pulse coupled neural network. Pattern Recogn Lett 28(9):1123\u20131132","journal-title":"Pattern Recogn Lett"},{"issue":"2","key":"19470_CR2","first-page":"262","volume":"5","author":"Y Yang","year":"2019","unstructured":"Yang Y, Nie Z, Huang S, Lin P, Wu J (2019) Multi-level features convolutional neural network for multi-focus image fusion. IEEE Trans Comput Imag 5(2):262\u2013273","journal-title":"IEEE Trans Comput Imag"},{"key":"19470_CR3","unstructured":"Aminnaji M, Aghagolzadeh A (2018) Multi-focus image fusion in dct domain using variance and energy of laplacian and correlation coefficient for visual sensor networks 56(2):233\u2013250"},{"issue":"5","key":"19470_CR4","doi-asserted-by":"publisher","first-page":"789","DOI":"10.1016\/j.compeleceng.2011.04.016","volume":"37","author":"M Haghighat","year":"2011","unstructured":"Haghighat M, Aghagolzadeh A, Seyedarabi H (2011) Multi-focus image fusion for visual sensor networks in dct domain. Comput Electr Eng 37(5):789\u2013797","journal-title":"Comput Electr Eng"},{"issue":"9","key":"19470_CR5","first-page":"4819","volume":"44","author":"X Zhang","year":"2021","unstructured":"Zhang X (2021) Deep learning-based multi-focus image fusion: a survey and a comparative study. IEEE Trans Pattern Anal Mach Intell 44(9):4819\u20134838","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"19470_CR6","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. Springer International Publishing, pp 234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"19470_CR7","doi-asserted-by":"crossref","unstructured":"Al-Mualla M, Ebrahim B, David H, Paul (2017) Perceptual image fusion using wavelets. IEEE Trans Image Process 26(3):1076\u20131088","DOI":"10.1109\/TIP.2016.2633863"},{"issue":"4","key":"19470_CR8","doi-asserted-by":"publisher","first-page":"884","DOI":"10.1109\/TIM.2009.2026612","volume":"59","author":"B Yang","year":"2010","unstructured":"Yang B, Li S (2010) Multifocus image fusion and restoration with sparse representation. IEEE Trans Instrument Measure 59(4):884\u2013892","journal-title":"IEEE Trans Instrument Measure"},{"issue":"5","key":"19470_CR9","doi-asserted-by":"publisher","first-page":"347","DOI":"10.1049\/iet-ipr.2014.0311","volume":"9","author":"Y Liu","year":"2014","unstructured":"Liu Y, Wang Z (2014) Simultaneous image fusion and denoising with adaptive sparse representation. Image Process Iet 9(5):347\u2013357","journal-title":"Image Process Iet"},{"key":"19470_CR10","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2014.10.004","volume":"25","author":"M Nejati","year":"2015","unstructured":"Nejati M, Samavi S, Shirani S (2015) Multi-focus image fusion using dictionary-based sparse representation. Inform Fusion 25:72\u201384","journal-title":"Inform Fusion"},{"issue":"5","key":"19470_CR11","doi-asserted-by":"publisher","first-page":"2045","DOI":"10.1109\/TIP.2016.2524212","volume":"25","author":"Q Zhang","year":"2016","unstructured":"Zhang Q, Levine M (2016) Robust multi-focus image fusion using multi-task sparse representation and spatial context. IEEE Trans Image Process 25(5):2045\u20132058","journal-title":"IEEE Trans Image Process"},{"issue":"3","key":"19470_CR12","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1016\/S1566-2535(01)00038-0","volume":"2","author":"S Li","year":"2001","unstructured":"Li S, Kwok JT, Wang Y (2001) Combination of images with diverse focuses using the spatial frequency. Inform Fusion 2(3):169\u2013176","journal-title":"Inform Fusion"},{"issue":"2","key":"19470_CR13","doi-asserted-by":"publisher","first-page":"136","DOI":"10.1016\/j.inffus.2012.01.007","volume":"14","author":"I De","year":"2013","unstructured":"De I (2013) Chanda B (2013) Multi-focus image fusion using a morphology-based focus measure in a quad-tree structure. Inform Fusion 14(2):136\u2013146","journal-title":"Inform Fusion"},{"key":"19470_CR14","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1016\/j.inffus.2014.05.003","volume":"22","author":"X Bai","year":"2015","unstructured":"Bai X, Zhang Y, Zhou F (2015) Xue B (2015) Quadtree-based multi-focus image fusion using a weighted focus-measure. Inform Fusion 22:105\u2013118","journal-title":"Inform Fusion"},{"key":"19470_CR15","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma J, Wei Y, Liang P, Chang L, Jiang J (2019) Fusiongan: a generative adversarial network for infrared and visible image fusion. Inform Fusion 48:11\u201326","journal-title":"Inform Fusion"},{"key":"19470_CR16","doi-asserted-by":"crossref","unstructured":"Xiao Y, Guo Z, Veelaert P, Philips W (2022) Dmdn: degradation model-based deep network for multi-focus image fusion, 101:116554,","DOI":"10.1016\/j.image.2021.116554"},{"key":"19470_CR17","doi-asserted-by":"crossref","unstructured":"Li Y, Shen X, Chen H (2019) Pyramid pooling dense convolutional neural network for multi-focus image fusion. In: 2019 IEEE 6th international conference on Cloud Computing and Intelligence Systems (CCIS), pp 164\u2013168","DOI":"10.1109\/CCIS48116.2019.9073690"},{"key":"19470_CR18","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu Y, Chen X, Peng H, Wang Z (2017) Multi-focus image fusion with a deep convolutional neural network. Inform Fusion 36:191\u2013207","journal-title":"Inform Fusion"},{"issue":"11","key":"19470_CR19","doi-asserted-by":"publisher","first-page":"5793","DOI":"10.1007\/s00521-020-05358-9","volume":"33","author":"B Ma","year":"2021","unstructured":"Ma B, Zhu Y, Yin X, Ban X, Huang H, Mukeshimana M (2021) Sesf-fuse: an unsupervised deep model for multi-focus image fusion. Neural Comput Appl 33(11):5793\u20135804","journal-title":"Neural Comput Appl"},{"key":"19470_CR20","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"ZA Yu","year":"2020","unstructured":"Yu ZA, Yu LB, Peng SC, Han YA, Xz D, Li ZA (2020) Ifcnn: a general image fusion framework based on convolutional neural network. Inform Fusion 54:99\u2013118","journal-title":"Inform Fusion"},{"key":"19470_CR21","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1016\/j.inffus.2017.10.007","volume":"42","author":"Y Liu","year":"2018","unstructured":"Liu Y, Chen X, Wang Z, Wang ZJ, Ward RK, Wang X (2018) Deep learning for pixel-level image fusion: recent advances and future prospects. Inform Fusion 42:158\u2013173","journal-title":"Inform Fusion"},{"key":"19470_CR22","doi-asserted-by":"crossref","unstructured":"Yu L, Xun C, Cheng J, Hu P (2018) A medical image fusion method based on convolutional neural networks. In: 2017 20th international conference on information fusion (Fusion), pp 1\u20137","DOI":"10.23919\/ICIF.2017.8009769"},{"key":"19470_CR23","doi-asserted-by":"crossref","unstructured":"Wang Q, Wu B, Zhu P, Li P, Hu Q (2020) Eca-net: efficient channel attention for deep convolutional neural networks. In: 2020 IEEE\/CVF conference on Computer Vision and Pattern Recognition (CVPR), pp 11534\u201311542","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"19470_CR24","doi-asserted-by":"crossref","unstructured":"Zhang H, Zu K, Lu J, Zou Y, Meng D (2022) Epsanet: an efficient pyramid squeeze attention block on convolutional neural network, 1161\u20131177","DOI":"10.1007\/978-3-031-26313-2_33"},{"key":"19470_CR25","doi-asserted-by":"crossref","unstructured":"Nejati M (2016) Lytro multi-focus image dataset 25:72\u201384","DOI":"10.1016\/j.inffus.2014.10.004"},{"key":"19470_CR26","unstructured":"Xu S, Wei X, Zhang C, Liu J, Zhang J (2020) Mffw: a new dataset for multi-focus image fusion"},{"key":"19470_CR27","doi-asserted-by":"crossref","unstructured":"Lin TY, Maire M, Belongie S, Hays J, Zitnick CL (2014) Microsoft coco: common objects in context. Springer International Publishing, 740\u2013755","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"19470_CR28","unstructured":"Kingma D, Ba J (2014) Adam: a method for stochastic optimization. Computer Science"},{"key":"19470_CR29","unstructured":"Paszke A, Gross S, Chintala S, Chanan G, Yang E, Devito Z, Lin Z, Desmaison A, Antiga L, Lerer A (2019) Automatic differentiation in pytorch"},{"key":"19470_CR30","unstructured":"Jie H, Li S, Gang S, Albanie S (2018) Squeeze-and-excitation networks. IEEE Trans Pattern Anal Mach Intell, 7132\u20137141"},{"key":"19470_CR31","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee JY, Kweon IS (2018) Cbam: convolutional block attention module. Proceedings of the European conference on computer vision (ECCV), pp 3\u201319","DOI":"10.1007\/978-3-030-01234-2_1"},{"issue":"1","key":"19470_CR32","first-page":"1","volume":"2","author":"AV Jan","year":"2008","unstructured":"Jan AV (2008) Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J Appl Remote Sens 2(1):1\u201328","journal-title":"J Appl Remote Sens"},{"issue":"12","key":"19470_CR33","doi-asserted-by":"publisher","first-page":"1890","DOI":"10.1016\/j.aeue.2015.09.004","volume":"69","author":"V Aslantas","year":"2015","unstructured":"Aslantas V, Bendes E (2015) A new image quality metric for image fusion: the sum of the correlations of differences. Aeu Int J Electron Commun 69(12):1890\u20131892","journal-title":"Aeu Int J Electron Commun"},{"key":"19470_CR34","unstructured":"Adams R, Horner M, Williams H (2011) Average gradient"},{"issue":"12","key":"19470_CR35","doi-asserted-by":"publisher","first-page":"2959","DOI":"10.1109\/26.477498","volume":"43","author":"AM Eskicioglu","year":"1995","unstructured":"Eskicioglu AM, Fisher PS (1995) Image quality measures and their performance. IEEE Trans Commun 43(12):2959\u20132965","journal-title":"IEEE Trans Commun"},{"key":"19470_CR36","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1016\/j.inffus.2014.05.004","volume":"23","author":"Y Liu","year":"2015","unstructured":"Liu Y, Liu S, Wang Z (2015) Multi-focus image fusion with dense SIFT. Inform Fusion 23:139\u2013155","journal-title":"Inform Fusion"},{"key":"19470_CR37","doi-asserted-by":"publisher","first-page":"201","DOI":"10.1016\/j.inffus.2019.02.003","volume":"51","author":"M Amin-Naji","year":"2019","unstructured":"Amin-Naji M, Aghagolzadeh A, Ezoji M (2019) Ensemble of cnn for multi-focus image fusion. int J Inform Fusion 51:201\u2013214","journal-title":"int J Inform Fusion"},{"key":"19470_CR38","doi-asserted-by":"publisher","first-page":"114385","DOI":"10.1109\/ACCESS.2019.2935006","volume":"7","author":"R Lai","year":"2019","unstructured":"Lai R, Li Y, Guan J, Xiong A (2019) Multi-scale visual attention deep convolutional neural network for multi-focus image fusion. IEEE Access 7:114385\u2013114399","journal-title":"IEEE Access"},{"key":"19470_CR39","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1016\/j.inffus.2013.11.005","volume":"20","author":"Z Zhou","year":"2014","unstructured":"Zhou Z, Li S, Wang B (2014) Multi-scale weighted gradient-based fusion for multi-focus images. Inform Fusion 20:60\u201372","journal-title":"Inform Fusion"},{"issue":"2","key":"19470_CR40","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1016\/j.inffus.2011.07.001","volume":"14","author":"S Li","year":"2013","unstructured":"Li S, Kang X, Hu J, Yang B (2013) Image matting for fusion of multi-focus images in dynamic scenes. Inform Fusion 14(2):147\u2013162","journal-title":"Inform Fusion"},{"key":"19470_CR41","first-page":"35","volume":"72","author":"X Qiu","year":"2019","unstructured":"Qiu X, Li M, Zhang L, Yuan X (2019) Guided filter-based multi-focus image fusion through focus region detection. Signal Process: Image Commun 72:35\u201346","journal-title":"Signal Process: Image Commun"},{"key":"19470_CR42","doi-asserted-by":"crossref","unstructured":"Zhang H, Xu H, Xiao Y, Guo X, Ma J (2020) Rethinking the image fusion: a fast unified image fusion network based on proportional maintenance of gradient and intensity. In: Proceedings of the AAAI conference on artificial intelligence, 34(7):12797\u201312804","DOI":"10.1609\/aaai.v34i07.6975"},{"key":"19470_CR43","doi-asserted-by":"publisher","first-page":"26316","DOI":"10.1109\/ACCESS.2020.2971137","volume":"8","author":"H Xu","year":"2020","unstructured":"Xu H, Fan F, Zhang H, Le Z, Huang J (2020) A deep model for multi-focus image fusion based on gradients and connected regions. IEEE Access 8:26316\u201326327","journal-title":"IEEE Access"},{"key":"19470_CR44","doi-asserted-by":"publisher","first-page":"4816","DOI":"10.1109\/TIP.2020.2976190","volume":"29","author":"J Li","year":"2020","unstructured":"Li J, Guo X, Lu G, Zhang B, Xu Y, Wu F, Zhang D (2020) Drpl: deep regression pair learning for multi-focus image fusion. IEEE Trans Image Process 29:4816\u20134831","journal-title":"IEEE Trans Image Process"},{"key":"19470_CR45","doi-asserted-by":"publisher","first-page":"102353","DOI":"10.1016\/j.displa.2022.102353","volume":"76","author":"P Wu","year":"2023","unstructured":"Wu P, Jiang L, Hua Z, Li J (2023) Multi-focus image fusion: transformer and shallow feature attention matters. Displays 76:102353","journal-title":"Displays"},{"issue":"5","key":"19470_CR46","doi-asserted-by":"publisher","first-page":"101978","DOI":"10.1016\/j.asej.2022.101978","volume":"14","author":"X Wang","year":"2023","unstructured":"Wang X, Hua Z, Li J (2023) Multi-focus image fusion framework based on transformer and feedback mechanism. Ain Shams Eng J 14(5):101978","journal-title":"Ain Shams Eng J"},{"key":"19470_CR47","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1016\/j.inffus.2022.11.014","volume":"92","author":"X Hu","year":"2023","unstructured":"Hu X, Jiang J, Liu X, Ma JZMFF (2023) Zero-shot multi-focus image fusion. Inform Fusion 92:127\u2013138","journal-title":"Inform Fusion"},{"issue":"7","key":"19470_CR48","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma J, Tang L, Fan F, Huang J, Mei X, Ma Y (2022) SwinFusion: cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA J Automatica Sinica 9(7):1200\u20131217","journal-title":"IEEE\/CAA J Automatica Sinica"},{"key":"19470_CR49","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2021.3118090","volume":"70","author":"Y Liu","year":"2021","unstructured":"Liu Y, Wang L, Cheng J, Chen X (2021) Multiscale feature interactive network for multifocus image fusion. IEEE Trans Instrum Meas 70:1\u201316","journal-title":"IEEE Trans Instrum Meas"},{"key":"19470_CR50","doi-asserted-by":"publisher","first-page":"174574","DOI":"10.1016\/j.eswa.2021.114574","volume":"171","author":"Z Wang","year":"2021","unstructured":"Wang Z, Li X, Duan H, Su Y, Zhang X, Guan X (2021) Medical image fusion based on convolutional neural networks and non-subsampled contourlet transform. Expert Syst Appl 171:174574","journal-title":"Expert Syst Appl"},{"key":"19470_CR51","doi-asserted-by":"publisher","first-page":"34483","DOI":"10.1007\/s11042-019-08070-6","volume":"78","author":"Z Wang","year":"2019","unstructured":"Wang Z, Li X, Duan H, Zhang X, Wang H (2019) Multifocus image fusion using convolutional neural networks in the discrete wavelet transform domain. Multimed Tool Appl 78:34483\u201334512","journal-title":"Multimed Tool Appl"},{"key":"19470_CR52","doi-asserted-by":"publisher","first-page":"4527","DOI":"10.1109\/TIP.2022.3184250","volume":"31","author":"Z Wang","year":"2022","unstructured":"Wang Z, Li X, Duan H, Zhang X (2022) A self-supervised residual feature learning model for multifocus image fusion. IEEE Trans Image Process 31:4527\u20134542","journal-title":"IEEE Trans Image Process"},{"issue":"10","key":"19470_CR53","doi-asserted-by":"publisher","first-page":"2529","DOI":"10.1007\/s11263-023-01806-w","volume":"131","author":"Z Wang","year":"2023","unstructured":"Wang Z, Li X, Zhao L, Duan H, Wang S, Liu H, Zhang X (2023) When multi-focus image fusion networks meet traditional edge-preservation technology. Int J Comput Vision 131(10):2529\u20132552","journal-title":"Int J Comput Vision"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19470-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-19470-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19470-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T10:29:19Z","timestamp":1747996159000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-19470-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,15]]},"references-count":53,"journal-issue":{"issue":"15","published-online":{"date-parts":[[2025,5]]}},"alternative-id":["19470"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-19470-8","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,6,15]]},"assertion":[{"value":"17 November 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 May 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 May 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 June 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest. The funding sponsors had no role in the study.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}}]}}