{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,12]],"date-time":"2026-02-12T15:23:28Z","timestamp":1770909808460,"version":"3.50.1"},"reference-count":31,"publisher":"Springer Science and Business Media LLC","issue":"22","license":[{"start":{"date-parts":[[2023,9,16]],"date-time":"2023-09-16T00:00:00Z","timestamp":1694822400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,9,16]],"date-time":"2023-09-16T00:00:00Z","timestamp":1694822400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62203224"],"award-info":[{"award-number":["62203224"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shanghai Special Plan for Local Colleges and Universities for Capacity Building","award":["22010501300"],"award-info":[{"award-number":["22010501300"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2023,11]]},"DOI":"10.1007\/s10489-023-04933-6","type":"journal-article","created":{"date-parts":[[2023,9,16]],"date-time":"2023-09-16T03:24:53Z","timestamp":1694834693000},"page":"27637-27654","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["ICWGAN-GP: an image fusion method based on infrared compensator and wasserstein generative adversarial network with gradient penalty"],"prefix":"10.1007","volume":"53","author":[{"given":"Xiao","family":"Wang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4143-828X","authenticated-orcid":false,"given":"Gang","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Lili","family":"Tang","sequence":"additional","affiliation":[]},{"given":"Durga Prasad","family":"Bavirisetti","sequence":"additional","affiliation":[]},{"given":"Gang","family":"Xiao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,16]]},"reference":[{"key":"4933_CR1","doi-asserted-by":"publisher","unstructured":"Ma J, Ma Y, Li C (2019) Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 45, 153\u2013178 (2019-01-01). https:\/\/doi.org\/10.1016\/j.inffus.2018.02.004. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253517307972","DOI":"10.1016\/j.inffus.2018.02.004"},{"issue":"108","key":"4933_CR2","first-page":"108","volume":"186","author":"L Ren","year":"2021","unstructured":"Ren L, Pan Z, Cao J, Zhang H, Wang H (2021) Infrared and visible image fusion based on edge-preserving guided filter and infrared feature decomposition. Signal Process. 186(108):108","journal-title":"Signal Process."},{"key":"4933_CR3","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen J, Li X, Luo L, Mei X, Ma J (2020) Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf Sci 508:64\u201378","journal-title":"Inf Sci"},{"key":"4933_CR4","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1016\/j.ins.2020.03.009","volume":"523","author":"H Li","year":"2020","unstructured":"Li H, He X, Yu Z, Luo J (2020) Noise-robust image fusion with low-rank sparse decomposition guided by external patch prior. Inf Sci 523:14\u201337","journal-title":"Inf Sci"},{"key":"4933_CR5","doi-asserted-by":"publisher","unstructured":"Chen J, Wu K, Cheng Z, Luo L (2021) A saliency-based multiscale approach for infrared and visible image fusion. Signal Process. (182): 107,936 https:\/\/doi.org\/10.1016\/j.sigpro.2020.107936. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0165168420304801","DOI":"10.1016\/j.sigpro.2020.107936"},{"issue":"5","key":"4933_CR6","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2019","unstructured":"Li H, Wu XJ (2019) DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans. Image Process. 28(5):2614\u20132623. https:\/\/doi.org\/10.1109\/TIP.2018.2887342","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"4933_CR7","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","volume":"69","author":"H Li","year":"2020","unstructured":"Li H, Wu XJ, Durrani T (2020) NestFuse: An Infrared and Visible Image Fusion Architecture Based on Nest Connection and Spatial\/Channel Attention Models. IEEE Trans Instrum Meas 69(12):9645\u20139656. https:\/\/doi.org\/10.1109\/TIM.2020.3005230","journal-title":"IEEE Trans Instrum Meas"},{"key":"4933_CR8","doi-asserted-by":"publisher","unstructured":"Li H, Wu XJ, Kittler J (2021) RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf Fusion 73:72\u201386. https:\/\/doi.org\/10.1016\/j.inffus.2021.02.023. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253521000440","DOI":"10.1016\/j.inffus.2021.02.023"},{"key":"4933_CR9","doi-asserted-by":"crossref","unstructured":"Vs V, Valanarasu JMJ, Oza P, Patel VM (2022) Image fusion transformer. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 3566\u20133570. IEEE","DOI":"10.1109\/ICIP46576.2022.9897280"},{"issue":"7","key":"4933_CR10","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma J, Tang L, Fan F, Huang J, Mei X, Ma Y (2022) Swinfusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA J Automatica Sinica 9(7):1200\u20131217","journal-title":"IEEE\/CAA J Automatica Sinica"},{"key":"4933_CR11","doi-asserted-by":"publisher","unstructured":"Ma J, Yu W, Liang P, Li C, Jiang J (2019) FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf Fusion 48:11\u201326. https:\/\/doi.org\/10.1016\/j.inffus.2018.09.004. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253518301143","DOI":"10.1016\/j.inffus.2018.09.004"},{"key":"4933_CR12","doi-asserted-by":"publisher","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","volume":"29","author":"J Ma","year":"2020","unstructured":"Ma J, Xu H, Jiang J, Mei X, Zhang XP (2020) DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion. IEEE Trans. Image Process. 29:4980\u20134995. https:\/\/doi.org\/10.1109\/TIP.2020.2977573","journal-title":"IEEE Trans. Image Process."},{"key":"4933_CR13","doi-asserted-by":"publisher","unstructured":"Ma J, Liang P, Yu W, Chen C, Guo X, Wu J, Jiang J (2020) Infrared and visible image fusion via detail preserving adversarial learning. Inf Fusion 54:85\u201398. https:\/\/doi.org\/10.1016\/j.inffus.2019.07.005. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253519300314","DOI":"10.1016\/j.inffus.2019.07.005"},{"key":"4933_CR14","doi-asserted-by":"publisher","unstructured":"Zhang Y, Zhang L, Bai X, Zhang L (2017) Infrared and visual image fusion through infrared feature extraction and visual information preservation. Infrared Phys. Technol 83:227\u2013237. https:\/\/doi.org\/10.1016\/j.infrared.2017.05.007. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1350449517300725","DOI":"10.1016\/j.infrared.2017.05.007"},{"key":"4933_CR15","unstructured":"Arjovsky M, Chintala S, Bottou L (2017) Wasserstein Generative Adversarial Networks. In: Proceedings of the 34th International Conference on Machine Learning, PMLR pp. 214\u2013223. https:\/\/proceedings.mlr.press\/v70\/arjovsky17a.html"},{"key":"4933_CR16","unstructured":"Gulrajani I, Ahmed F, Arjovsky M, Dumoulin V, Courville AC (2017) Improved Training of Wasserstein GANs. In: Advances in Neural Information Processing Systems, vol.\u00a030. Curran Associates, Inc.. https:\/\/proceedings.neurips.cc\/paper\/2017\/hash\/892c3b1c6dccd52936e27cbd0ff683d6-Abstract.html"},{"issue":"10","key":"4933_CR17","doi-asserted-by":"publisher","first-page":"1786","DOI":"10.1109\/TPAMI.2007.70830","volume":"30","author":"L Zhang","year":"2008","unstructured":"Zhang L (2008) In situ image segmentation using the convexity of illumination distribution of the light sources. IEEE transactions on pattern analysis and machine intelligence 30(10):1786\u20131799","journal-title":"IEEE transactions on pattern analysis and machine intelligence"},{"issue":"5","key":"4933_CR18","doi-asserted-by":"publisher","first-page":"1193","DOI":"10.1007\/s11760-013-0556-9","volume":"9","author":"BK Shreyamsha Kumar","year":"2015","unstructured":"Shreyamsha Kumar BK (2015) Image fusion based on pixel significance using cross bilateral filter. Signal, Image and Video Processing 9(5):1193\u20131204. https:\/\/doi.org\/10.1007\/s11760-013-0556-9","journal-title":"Signal, Image and Video Processing"},{"key":"4933_CR19","doi-asserted-by":"publisher","unstructured":"Ma J, Chen C, Li C, Huang J (2016) Infrared and visible image fusion via gradient transfer and total variation minimization. Inf Fusion 31:100\u2013109. https:\/\/doi.org\/10.1016\/j.inffus.2016.02.001. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S156625351630001X","DOI":"10.1016\/j.inffus.2016.02.001"},{"key":"4933_CR20","doi-asserted-by":"publisher","unstructured":"Bavirisetti DP, Xiao G, Liu G (2017) Multi-sensor image fusion based on fourth order partial differential equations. In: 2017 20th International Conference on Information Fusion (Fusion), pp. 1\u20139. https:\/\/doi.org\/10.23919\/ICIF.2017.8009719","DOI":"10.23919\/ICIF.2017.8009719"},{"issue":"1","key":"4933_CR21","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2022","unstructured":"Xu H, Ma J, Jiang J, Guo X, Ling H (2022) U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal Mach Intell 44(1):502\u2013518. https:\/\/doi.org\/10.1109\/TPAMI.2020.3012548","journal-title":"IEEE Trans. Pattern Anal Mach Intell"},{"key":"4933_CR22","doi-asserted-by":"publisher","first-page":"4733","DOI":"10.1109\/TIP.2020.2975984","volume":"29","author":"H Li","year":"2020","unstructured":"Li H, Wu XJ, Kittler J (2020) MDLatLRR: A Novel Decomposition Method for Infrared and Visible Image Fusion. IEEE Trans. Image Process 29:4733\u20134746. https:\/\/doi.org\/10.1109\/TIP.2020.2975984","journal-title":"IEEE Trans. Image Process"},{"key":"4933_CR23","doi-asserted-by":"publisher","unstructured":"Ma J, Zhang H, Shao Z, Liang P, Xu H (2021) GANMcC: A Generative Adversarial Network With Multiclassification Constraints for Infrared and Visible Image Fusion. IEEE Trans Instrum Meas 70:1\u201314. https:\/\/doi.org\/10.1109\/TIM.2020.3038013","DOI":"10.1109\/TIM.2020.3038013"},{"key":"4933_CR24","doi-asserted-by":"publisher","unstructured":"Meher B, Agrawal S, Panda R, Abraham A (2019) A survey on region based image fusion methods. Inf Fusion 48, 119\u2013132 (2019-08-01). https:\/\/doi.org\/10.1016\/j.inffus.2018.07.010. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253517307583","DOI":"10.1016\/j.inffus.2018.07.010"},{"key":"4933_CR25","doi-asserted-by":"publisher","unstructured":"Han Y, Cai Y, Cao Y, Xu X (2013) A new image fusion performance metric based on visual information fidelity. Inf Fusion 14(2):127\u2013135. https:\/\/doi.org\/10.1016\/j.inffus.2011.08.002. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S156625351100056X","DOI":"10.1016\/j.inffus.2011.08.002"},{"key":"4933_CR26","unstructured":"Deshmukh M, Bhosale U, Deshmukh M, Engg SCO, Mumbai N, Bhosale U (2010) Image Fusion and Image Quality Assessment of Fused Images. International Journal of Image Processing (IJIP"},{"issue":"4","key":"4933_CR27","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang Z, Bovik A, Sheikh H (2004) Simoncelli E (2004) Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 13(4):600\u2013612. https:\/\/doi.org\/10.1109\/TIP.2003.819861","journal-title":"IEEE Trans. Image Process."},{"key":"4933_CR28","doi-asserted-by":"publisher","unstructured":"Aslantas V (2015) Bendes E (2015) A new image quality metric for image fusion: The sum of the correlations of differences. AEU - International J Electron Commun 69(12):1890\u20131896.\u00a0https:\/\/doi.org\/10.1016\/j.aeue.2015.09.004. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1434841115002691","DOI":"10.1016\/j.aeue.2015.09.004"},{"key":"4933_CR29","doi-asserted-by":"crossref","unstructured":"Haghighat M, Razian MA (2014) Fast-fmi: non-reference image fusion metric.In: 2014 IEEE 8th International Conference on Application of Information and Communication Technologies (AICT), IEEE pp. 1\u20133","DOI":"10.1109\/ICAICT.2014.7036000"},{"key":"4933_CR30","doi-asserted-by":"publisher","unstructured":"Zhang X, Ye P, Leung H, Gong K, Xiao G (2020) Object fusion tracking based on visible and infrared images: A comprehensive review. Inf Fusion 63:166\u2013187. https:\/\/doi.org\/10.1016\/j.inffus.2020.05.002. https:\/\/www.sciencedirect.com\/science\/article\/pii\/S1566253520302657","DOI":"10.1016\/j.inffus.2020.05.002"},{"key":"4933_CR31","doi-asserted-by":"crossref","unstructured":"Yang T, Xu P, Hu R, Chai H, Chan AB (2020) ROAM: Recurrently Optimizing Tracking Model. pp. 6718\u20136727. https:\/\/openaccess.thecvf.com\/content_CVPR_2020\/html\/Yang_ROAM_Recurrently_Optimizing_Tracking_Model_CVPR_2020_paper.html","DOI":"10.1109\/CVPR42600.2020.00675"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-023-04933-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-023-04933-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-023-04933-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,10,25]],"date-time":"2023-10-25T23:18:45Z","timestamp":1698275925000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-023-04933-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9,16]]},"references-count":31,"journal-issue":{"issue":"22","published-print":{"date-parts":[[2023,11]]}},"alternative-id":["4933"],"URL":"https:\/\/doi.org\/10.1007\/s10489-023-04933-6","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,9,16]]},"assertion":[{"value":"31 July 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 September 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We declare that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work, there is no professional or other personal interest of any nature or kind in any product, service or company that could be construed as influencing the position presented in the manuscript entitled \"Manuscript\".","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}]}}