{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,30]],"date-time":"2025-09-30T04:39:25Z","timestamp":1759207165750,"version":"3.37.3"},"reference-count":31,"publisher":"Springer Science and Business Media LLC","issue":"13","license":[{"start":{"date-parts":[[2023,9,30]],"date-time":"2023-09-30T00:00:00Z","timestamp":1696032000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,9,30]],"date-time":"2023-09-30T00:00:00Z","timestamp":1696032000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100017700","name":"Henan Provincial Science and Technology Research Project","doi-asserted-by":"publisher","award":["No.222102210309"],"award-info":[{"award-number":["No.222102210309"]}],"id":[{"id":"10.13039\/501100017700","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100014718","name":"Innovative Research Group Project of the National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["Nos. 62106067","62106068"],"award-info":[{"award-number":["Nos. 62106067","62106068"]}],"id":[{"id":"10.13039\/100014718","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100006407","name":"Natural Science Foundation of Henan Province","doi-asserted-by":"publisher","award":["No.21A520010"],"award-info":[{"award-number":["No.21A520010"]}],"id":[{"id":"10.13039\/501100006407","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100018607","name":"Henan Province Science and Technology Innovation Talent Program","doi-asserted-by":"publisher","award":["2021ZKCJ14"],"award-info":[{"award-number":["2021ZKCJ14"]}],"id":[{"id":"10.13039\/501100018607","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-023-16978-3","type":"journal-article","created":{"date-parts":[[2023,9,30]],"date-time":"2023-09-30T07:02:16Z","timestamp":1696057336000},"page":"37505-37522","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["TCGFusion: a network for PET-MRI fusion based on GAN and transformer"],"prefix":"10.1007","volume":"83","author":[{"given":"Chao","family":"Fan","sequence":"first","affiliation":[]},{"given":"Zhixiang","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Hao","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Xiao","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,30]]},"reference":[{"key":"16978_CR1","unstructured":"Ambati LS, El-Gayar OF, Nawar N (2021) Design principles for multiple sclerosis mobile self-management applications:\u00a0a patient-centric perspective[C]\/\/AMCIS"},{"issue":"12","key":"16978_CR2","doi-asserted-by":"publisher","first-page":"11742","DOI":"10.1002\/int.23061","volume":"37","author":"IF Zamzami","year":"2022","unstructured":"Zamzami IF, Pathoee K, Gupta BB et al (2022) Machine learning algorithms for smart and intelligent healthcare system in Society 5.0. Int J Intell Syst 37(12):11742\u201311763","journal-title":"Int J Intell Syst"},{"issue":"1","key":"16978_CR3","first-page":"1","volume":"14","author":"A Gaurav","year":"2022","unstructured":"Gaurav A, Psannis K, Perakovi\u0107 D (2022) Security of cloud-based medical internet of things (miots): A survey. Int J Software Sci Comput Intell (IJSSCI) 14(1):1\u201316","journal-title":"Int J Software Sci Comput Intell (IJSSCI)"},{"key":"16978_CR4","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.05.004","volume":"33","author":"S Li","year":"2017","unstructured":"Li S, Kang X, Fang L et al (2017) Pixel-level image fusion: a survey of the state of the art. Inf Fusion 33:100\u2013112","journal-title":"Inf Fusion"},{"issue":"23","key":"16978_CR5","doi-asserted-by":"publisher","first-page":"4184","DOI":"10.1016\/j.ijleo.2015.08.118","volume":"126","author":"W Zhi-She","year":"2015","unstructured":"Zhi-She W, Feng-Bao Y, Zhi-Hao P et al (2015) Multi-sensor image enhanced fusion algorithm based on NSST and top-hat transformation. Optik - Int J Light Electron Optics 126(23):4184\u20134190","journal-title":"Optik - Int J Light Electron Optics"},{"key":"16978_CR6","doi-asserted-by":"crossref","unstructured":"Wang Z, Xu J, Jiang X et al (2020) Infrared and visible image fusion via hybrid decomposition of NSCT and morphological sequential toggle\u00a0operator[J]. Optik\u00a0201:163497","DOI":"10.1016\/j.ijleo.2019.163497"},{"key":"16978_CR7","doi-asserted-by":"crossref","unstructured":"Zhang Q, Liu Y, Blum RS et al (2018) Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: a review[J]. Information Fusion\u00a040:57\u201375","DOI":"10.1016\/j.inffus.2017.05.006"},{"issue":"2","key":"16978_CR8","doi-asserted-by":"publisher","first-page":"220","DOI":"10.1109\/LSP.2014.2354534","volume":"22","author":"L Cao","year":"2014","unstructured":"Cao L, Jin L, Tao H et al (2014) Multi-Focus Image Fusion Based on Spatial Frequency in Discrete Cosine Transform Domain. IEEE Signal Process Lett 22(2):220\u2013224","journal-title":"IEEE Signal Process Lett"},{"key":"16978_CR9","doi-asserted-by":"crossref","unstructured":"Hu P, Yang F, Wei H et al (2019) A multi-algorithm block fusion method based on set-valued mapping for dual-modal infrared\u00a0images[J].\u00a0Infrared Phys Technol\u00a0102:102977","DOI":"10.1016\/j.infrared.2019.102977"},{"key":"16978_CR10","doi-asserted-by":"crossref","unstructured":"Li Z, Hu HM, Zhang W et al (2020) Spectrum characteristics preserved visible and near-infrared image fusion algorithm[J]. IEEE Trans Multimedia\u00a023:306\u2013319","DOI":"10.1109\/TMM.2020.2978640"},{"key":"16978_CR11","doi-asserted-by":"crossref","unstructured":"Nie R, Ma C, Cao J et al (2021) A total variation with joint norms for infrared and visible image fusion[J]. IEEE Trans Multimedia 24:1460\u20131472","DOI":"10.1109\/TMM.2021.3065496"},{"issue":"9","key":"16978_CR12","doi-asserted-by":"publisher","first-page":"1797","DOI":"10.1007\/s00371-019-01774-8","volume":"36","author":"S Zhang","year":"2020","unstructured":"Zhang S, He F (2020) DRCDN: Learning deep residual convolutional dehazing networks. Vis Comput 36(9):1797\u20131808","journal-title":"Vis Comput"},{"key":"16978_CR13","doi-asserted-by":"publisher","unstructured":"Zhang Y, Liu Y, Sun P, et al. IFCNN: A General Image Fusion Framework Based on Convolutional Neural Network. information Fusion 54. https:\/\/doi.org\/10.1016\/j.infus.2019.07.011","DOI":"10.1016\/j.infus.2019.07.011"},{"key":"16978_CR14","doi-asserted-by":"crossref","unstructured":"Xu H, Ma J, Le Z et al (2020)\u00a0Fusiondn: A unified densely connected network for image fusion[C]\/\/Proceedings of the\u00a0AAAI conference on artificial intelligence 34(07):12484\u201312491","DOI":"10.1609\/aaai.v34i07.6936"},{"key":"16978_CR15","doi-asserted-by":"publisher","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","volume":"29","author":"J Ma","year":"2020","unstructured":"Ma J, Xu H, Jiang J et al (2020) DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion. IEEE Trans Image Process 29:4980\u20134995. https:\/\/doi.org\/10.1109\/TIP.2020.2977573","journal-title":"IEEE Trans Image Process"},{"issue":"1","key":"16978_CR16","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2022","unstructured":"Xu H, Ma J, Jiang J et al (2022) U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans Pattern Anal Machine Intell 44(1):502\u2013518. https:\/\/doi.org\/10.1109\/TPAMI.2020.3012548","journal-title":"IEEE Trans Pattern Anal Machine Intell"},{"key":"16978_CR17","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1016\/j.inffus.2022.11.010","volume":"92","author":"C Cheng","year":"2023","unstructured":"Cheng C, Xu T, Wu XJ (2023) MUFusion: A general unsupervised image fusion network based on memory unit. Inf Fusion 92:80\u201392","journal-title":"Inf Fusion"},{"key":"16978_CR18","doi-asserted-by":"publisher","first-page":"55145","DOI":"10.1109\/ACCESS.2020.2982016","volume":"8","author":"J Huang","year":"2020","unstructured":"Huang J, Le Z, Ma Y et al (2020) MGMDcGAN: Medical Image Fusion Using Multi-Generator Multi-Discriminator Conditional Generative Adversarial Network. IEEE Access 8:55145\u201355157. https:\/\/doi.org\/10.1109\/ACCESS.2020.2982016","journal-title":"IEEE Access"},{"key":"16978_CR19","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1016\/j.inffus.2021.06.001","volume":"76","author":"H Xu","year":"2021","unstructured":"Xu H, Ma J (2021) EMFusion: An unsupervised enhanced medical image fusion network. Inf Fusion 76:177\u2013186","journal-title":"Inf Fusion"},{"key":"16978_CR20","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A et al (2020)\u00a0An image is worth 16x16 words: Transformers for image\u00a0recognition at scale[J]. arXiv:2010.11929"},{"key":"16978_CR21","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y et al (2021) Swin transformer: hierarchical vision transformer using shifted\u00a0windows[C]\/\/Proceedings of the IEEE\/CVF international conference on computer vision, pp 10012\u201310022","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"16978_CR22","doi-asserted-by":"crossref","unstructured":"Rao D, Xu T, Wu XJ (2023) Tgfuse: an Infrared and visible image fusion approach based on transformer and generative adversarial network[J]. IEEE Transactions on Image Processing","DOI":"10.1109\/TIP.2023.3273451"},{"key":"16978_CR23","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang L, Yuan J, Ma J (2022) Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inf Fusion 82:28\u201342","journal-title":"Inf Fusion"},{"key":"16978_CR24","doi-asserted-by":"crossref","unstructured":"Xydeas CS, Petrovic VS (2000) Objective pixel-level image fusion performance measure[C]\/\/Sensor Fusion: Architectures, Algorithms, and Applications IV. SPIE 4051:89\u201398","DOI":"10.1117\/12.381668"},{"issue":"11","key":"16978_CR25","doi-asserted-by":"publisher","first-page":"3345","DOI":"10.1109\/TIP.2015.2442920","volume":"24","author":"K Ma","year":"2015","unstructured":"Ma K, Zeng K, Wang Z (2015) Perceptual quality assessment for multi-exposure image fusion. IEEE Trans Image Process 24(11):3345\u20133356","journal-title":"IEEE Trans Image Process"},{"issue":"4","key":"16978_CR26","doi-asserted-by":"publisher","first-page":"308","DOI":"10.1049\/el:20000267","volume":"36","author":"C Xydeas","year":"2000","unstructured":"Xydeas C, Petrovic V (2000) Objective image fusion performance measure. Electron Lett 36(4):308\u2013309","journal-title":"Electron Lett"},{"issue":"3","key":"16978_CR27","doi-asserted-by":"publisher","first-page":"660","DOI":"10.1109\/TBC.2008.2000733","volume":"54","author":"S Winkler","year":"2008","unstructured":"Winkler S, Mohandas P (2008) The Evolution of Video Quality Measurement: From PSNR to Hybrid Metrics. IEEE Trans Broadcast 54(3):660\u2013668. https:\/\/doi.org\/10.1109\/TBC.2008.2000733","journal-title":"IEEE Trans Broadcast"},{"key":"16978_CR28","doi-asserted-by":"crossref","unstructured":"G. Qu, D. Zhang, P. Yan, \"Information measure for performance of image fusion,\" Electronics letters, vol. 38, no. 7, pp. 313\u2013315, 2002. 8","DOI":"10.1049\/el:20020212"},{"key":"16978_CR29","doi-asserted-by":"crossref","unstructured":"Haghighat M, Razian MA (2014)\u00a0Fast-FMI: Non-reference image fusion metric[C]\/\/2014 IEEE 8th International\u00a0Conference on Application of Information and Communication Technologies (AICT). IEEE pp\u00a01\u20133","DOI":"10.1109\/ICAICT.2014.7036000"},{"key":"16978_CR30","doi-asserted-by":"publisher","first-page":"584","DOI":"10.1109\/TCI.2021.3083965","volume":"7","author":"W Tang","year":"2021","unstructured":"Tang W, Liu Y, Cheng J et al (2021) Green Fluorescent Protein and Phase Contrast Image Fusion Via Detail Preserving Cross Network. IEEE Trans Comput Imaging 7:584\u2013597. https:\/\/doi.org\/10.1109\/TCI.2021.3083965","journal-title":"IEEE Trans Comput Imaging"},{"key":"16978_CR31","doi-asserted-by":"publisher","unstructured":"Wang LT, Hoover NE, Porte EH, et al. (1987) \"SSIM: A Software Levelized Compiled-Code Simulator.\" 24th ACM\/IEEE Design Automation Conference, pp. 2\u20138. doi: https:\/\/doi.org\/10.1109\/DAC.1987.203214","DOI":"10.1109\/DAC.1987.203214"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16978-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-16978-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16978-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,4,3]],"date-time":"2024-04-03T10:15:23Z","timestamp":1712139323000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-16978-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9,30]]},"references-count":31,"journal-issue":{"issue":"13","published-online":{"date-parts":[[2024,4]]}},"alternative-id":["16978"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-16978-3","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2023,9,30]]},"assertion":[{"value":"30 October 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 September 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 September 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 September 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article. All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}