{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,30]],"date-time":"2026-04-30T11:39:46Z","timestamp":1777549186897,"version":"3.51.4"},"reference-count":50,"publisher":"Springer Science and Business Media LLC","issue":"30","license":[{"start":{"date-parts":[[2023,8,13]],"date-time":"2023-08-13T00:00:00Z","timestamp":1691884800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,8,13]],"date-time":"2023-08-13T00:00:00Z","timestamp":1691884800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"The provincial major science and technology special plan project of Yunnan","award":["202202AD080003"],"award-info":[{"award-number":["202202AD080003"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62202416"],"award-info":[{"award-number":["62202416"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62162068"],"award-info":[{"award-number":["62162068"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2023,10]]},"DOI":"10.1007\/s00521-023-08916-z","type":"journal-article","created":{"date-parts":[[2023,8,13]],"date-time":"2023-08-13T04:01:10Z","timestamp":1691899270000},"page":"22511-22529","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Superpixel-based adaptive salient region analysis for infrared and visible image fusion"],"prefix":"10.1007","volume":"35","author":[{"given":"Chengzhou","family":"Li","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6207-9728","authenticated-orcid":false,"given":"Kangjian","family":"He","sequence":"additional","affiliation":[]},{"given":"Dan","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Dapeng","family":"Tao","sequence":"additional","affiliation":[]},{"given":"Xu","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Hongzhen","family":"Shi","sequence":"additional","affiliation":[]},{"given":"Wenxia","family":"Yin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,13]]},"reference":[{"key":"8916_CR1","first-page":"1","volume":"70","author":"K He","year":"2021","unstructured":"He K, Gong J, Xie L et al (2021) Regions preserving edge enhancement for multisensor-based medical image fusion. IEEE Trans Instrum Meas 70:1\u201313","journal-title":"IEEE Trans Instrum Meas"},{"key":"8916_CR2","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Xu D, et al (2022) Fidelity-driven optimization reconstruction and details preserving guided fusion for multi-modality medical image. IEEE Trans Multimedia","DOI":"10.1109\/TMM.2022.3185887"},{"key":"8916_CR3","doi-asserted-by":"publisher","first-page":"20831","DOI":"10.1007\/s00521-022-07559-w","volume":"34","author":"W Yin","year":"2022","unstructured":"Yin W, He K, Xu D et al (2022) Adaptive enhanced infrared and visible image fusion using hybrid decomposition and coupled dictionary. Neural Comput Appl 34:20831\u201320849","journal-title":"Neural Comput Appl"},{"key":"8916_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2022.104041","volume":"121","author":"W Yin","year":"2022","unstructured":"Yin W, He K, Xu D et al (2022) Significant target analysis and detail preserving based infrared and visible image fusion. Infrared Phys Technol 121:104041","journal-title":"Infrared Phys Technol"},{"key":"8916_CR5","first-page":"1","volume":"71","author":"Q Jiang","year":"2022","unstructured":"Jiang Q, Lee S, Zeng X et al (2022) A multifocus image fusion scheme based on similarity measure of transformed isosceles triangles between intuitionistic fuzzy sets. IEEE Trans Instrum Meas 71:1\u201315","journal-title":"IEEE Trans Instrum Meas"},{"key":"8916_CR6","doi-asserted-by":"publisher","first-page":"478","DOI":"10.1016\/j.infrared.2017.07.010","volume":"85","author":"X Jin","year":"2017","unstructured":"Jin X, Jiang Q, Yao S et al (2017) A survey of infrared and visual image fusion methods. Infrared Phys Technol 85:478\u2013501","journal-title":"Infrared Phys Technol"},{"key":"8916_CR7","doi-asserted-by":"crossref","unstructured":"Noori FM, Riegler M, Uddin MZ, Torresen J (2020) Human activity recognition from multiple sensors data using multi-fusion representations and CNNs. ACM Trans Multimedia Comput Commun Appl (TOMM) 16:1\u201319","DOI":"10.1145\/3377882"},{"key":"8916_CR8","doi-asserted-by":"publisher","first-page":"376","DOI":"10.1016\/j.inffus.2022.10.022","volume":"91","author":"Z Zhu","year":"2023","unstructured":"Zhu Z, He X, Qi G et al (2023) Brain tumor segmentation based on the fusion of deep semantics and edge information in multimodal MRI. Inf Fusion 91:376\u2013387. https:\/\/doi.org\/10.1016\/j.inffus.2022.10.022","journal-title":"Inf Fusion"},{"key":"8916_CR9","doi-asserted-by":"publisher","first-page":"182","DOI":"10.1016\/j.neucom.2016.11.051","volume":"226","author":"M Yin","year":"2017","unstructured":"Yin M, Duan P, Liu W, Liang X (2017) A novel infrared and visible image fusion algorithm based on shift-invariant dual-tree complex shearlet transform and sparse representation. Neurocomputing 226:182\u2013191","journal-title":"Neurocomputing"},{"key":"8916_CR10","doi-asserted-by":"publisher","first-page":"6133","DOI":"10.1007\/s00521-020-05387-4","volume":"33","author":"Z Yang","year":"2021","unstructured":"Yang Z, Chen Y, Le Z, Ma Y (2021) GANFuse: a novel multi-exposure image fusion method based on generative adversarial networks. Neural Comput Appl 33:6133\u20136145","journal-title":"Neural Comput Appl"},{"key":"8916_CR11","doi-asserted-by":"publisher","first-page":"8137","DOI":"10.1007\/s00521-022-06900-7","volume":"34","author":"S Singh","year":"2022","unstructured":"Singh S, Mittal N, Singh H (2022) A feature level image fusion for IR and visible image using mNMRA based segmentation. Neural Comput Appl 34:8137\u20138154","journal-title":"Neural Comput Appl"},{"key":"8916_CR12","doi-asserted-by":"publisher","first-page":"469","DOI":"10.1109\/34.3911","volume":"10","author":"N Nandhakumar","year":"1988","unstructured":"Nandhakumar N, Aggarwal JK (1988) Integrated analysis of thermal and visual images for scene interpretation. IEEE Trans Pattern Anal Mach Intell 10:469\u2013481","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"8916_CR13","doi-asserted-by":"crossref","unstructured":"Morris NJ, Avidan S, Matusik W, Pfister H (2007) Statistics of infrared images. In: 2007 IEEE conference on computer vision and pattern recognition. IEEE, New York, pp 1\u20137","DOI":"10.1109\/CVPR.2007.383003"},{"key":"8916_CR14","doi-asserted-by":"publisher","first-page":"7033","DOI":"10.1364\/AO.52.007033","volume":"52","author":"L Li","year":"2013","unstructured":"Li L, Li H, Dang E, Liu B (2013) Compressive sensing method for recognizing cat-eye effect targets. Appl Opt 52:7033\u20137039","journal-title":"Appl Opt"},{"key":"8916_CR15","doi-asserted-by":"publisher","first-page":"510","DOI":"10.1049\/el.2014.0180","volume":"50","author":"L Li","year":"2014","unstructured":"Li L, Li H, Li T, Gao F (2014) Infrared small target detection in compressive domain. Electron Lett 50:510\u2013512","journal-title":"Electron Lett"},{"key":"8916_CR16","doi-asserted-by":"publisher","first-page":"3422","DOI":"10.1109\/TCSVT.2019.2963721","volume":"30","author":"K Kansal","year":"2020","unstructured":"Kansal K, Subramanyam AV, Wang Z, Satoh S (2020) SDL: spectrum-disentangled representation learning for visible-infrared person re-identification. IEEE Trans Circuits Syst Video Technol 30:3422\u20133432","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"8916_CR17","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","volume":"82","author":"J Ma","year":"2017","unstructured":"Ma J, Zhou Z, Wang B, Zong H (2017) Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys Technol 82:8\u201317","journal-title":"Infrared Phys Technol"},{"key":"8916_CR18","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.02.001","volume":"31","author":"J Ma","year":"2016","unstructured":"Ma J, Chen C, Li C, Huang J (2016) Infrared and visible image fusion via gradient transfer and total variation minimization. Information Fusion 31:100\u2013109","journal-title":"Information Fusion"},{"key":"8916_CR19","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1109\/JSEN.2015.2478655","volume":"16","author":"DP Bavirisetti","year":"2015","unstructured":"Bavirisetti DP, Dhuli R (2015) Fusion of infrared and visible sensor images based on anisotropic diffusion and Karhunen-Loeve transform. IEEE Sens J 16:203\u2013209","journal-title":"IEEE Sens J"},{"key":"8916_CR20","doi-asserted-by":"publisher","first-page":"52","DOI":"10.1016\/j.infrared.2016.01.009","volume":"76","author":"DP Bavirisetti","year":"2016","unstructured":"Bavirisetti DP, Dhuli R (2016) Two-scale image fusion of visible and infrared images using saliency detection. Infrared Phys Technol 76:52\u201364","journal-title":"Infrared Phys Technol"},{"key":"8916_CR21","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","volume":"508","author":"J Chen","year":"2020","unstructured":"Chen J, Li X, Luo L et al (2020) Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf Sci 508:64\u201378","journal-title":"Inf Sci"},{"key":"8916_CR22","doi-asserted-by":"crossref","unstructured":"Burt PJ, Adelson EH (1987) The Laplacian pyramid as a compact image code. In: Readings in computer vision. Elsevier, Amsredam, pp 671\u2013679","DOI":"10.1016\/B978-0-08-051581-6.50065-9"},{"key":"8916_CR23","first-page":"1","volume":"2012","author":"Y Niu","year":"2012","unstructured":"Niu Y, Xu S, Wu L, Hu W (2012) Airborne infrared and visible image fusion for target perception based on target region segmentation and discrete wavelet transform. Math Problems Eng 2012:1\u201310","journal-title":"Math Problems Eng"},{"key":"8916_CR24","doi-asserted-by":"publisher","first-page":"2","DOI":"10.1016\/j.sigpro.2016.12.025","volume":"136","author":"P Fletcher","year":"2017","unstructured":"Fletcher P, Sangwine SJ (2017) The development of the quaternion wavelet transform. Signal Process 136:2\u201315","journal-title":"Signal Process"},{"key":"8916_CR25","doi-asserted-by":"publisher","first-page":"3089","DOI":"10.1109\/TIP.2006.877507","volume":"15","author":"AL Da Cunha","year":"2006","unstructured":"Da Cunha AL, Zhou J, Do MN (2006) The nonsubsampled contourlet transform: theory, design, and applications. IEEE Trans Image Process 15:3089\u20133101","journal-title":"IEEE Trans Image Process"},{"key":"8916_CR26","doi-asserted-by":"publisher","first-page":"3064","DOI":"10.1364\/AO.58.003064","volume":"58","author":"W Tan","year":"2019","unstructured":"Tan W, Zhou H, Song J et al (2019) Infrared and visible image perceptive fusion through multi-level Gaussian curvature filtering image decomposition. Appl Opt 58:3064\u20133073","journal-title":"Appl Opt"},{"key":"8916_CR27","doi-asserted-by":"publisher","unstructured":"Li H, Qi X, Xie W (2020) Fast infrared and visible image fusion with structural decomposition. Knowledge-Based Syst 204:106182. https:\/\/doi.org\/10.1016\/j.knosys.2020.106182","DOI":"10.1016\/j.knosys.2020.106182"},{"key":"8916_CR28","doi-asserted-by":"publisher","first-page":"210","DOI":"10.1109\/TPAMI.2008.79","volume":"31","author":"J Wright","year":"2008","unstructured":"Wright J, Yang AY, Ganesh A et al (2008) Robust face recognition via sparse representation. IEEE Trans Pattern Anal Mach Intell 31:210\u2013227","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"8916_CR29","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1109\/TPAMI.2012.88","volume":"35","author":"G Liu","year":"2012","unstructured":"Liu G, Lin Z, Yan S et al (2012) Robust recovery of subspace structures by low-rank representation. IEEE Trans Pattern Anal Mach Intell 35:171\u2013184","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"8916_CR30","doi-asserted-by":"publisher","first-page":"4733","DOI":"10.1109\/TIP.2020.2975984","volume":"29","author":"H Li","year":"2020","unstructured":"Li H, Wu X-J, Kittler J (2020) MDLatLRR: a novel decomposition method for infrared and visible image fusion. IEEE Trans Image Process 29:4733\u20134746","journal-title":"IEEE Trans Image Process"},{"key":"8916_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.optlaseng.2022.107078","volume":"156","author":"B Wang","year":"2022","unstructured":"Wang B, Zou Y, Zhang L et al (2022) Multimodal super-resolution reconstruction of infrared and visible images via deep learning. Opt Lasers Eng 156:107078","journal-title":"Opt Lasers Eng"},{"key":"8916_CR32","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu Y, Chen X, Peng H, Wang Z (2017) Multi-focus image fusion with a deep convolutional neural network. Inform Fusion 36:191\u2013207","journal-title":"Inform Fusion"},{"key":"8916_CR33","doi-asserted-by":"publisher","first-page":"1982","DOI":"10.1109\/TMM.2019.2895292","volume":"21","author":"X Guo","year":"2019","unstructured":"Guo X, Nie R, Cao J et al (2019) FuseGAN: learning to fuse multi-focus image via conditional generative adversarial network. IEEE Trans Multimedia 21:1982\u20131996","journal-title":"IEEE Trans Multimedia"},{"key":"8916_CR34","doi-asserted-by":"publisher","first-page":"2761","DOI":"10.1007\/s11263-021-01501-8","volume":"129","author":"H Zhang","year":"2021","unstructured":"Zhang H, Ma J (2021) SDNet: a versatile squeeze-and-decomposition network for real-time image fusion. Int J Comput Vis 129:2761\u20132785","journal-title":"Int J Comput Vis"},{"key":"8916_CR35","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2022","unstructured":"Xu H, Ma J, Jiang J et al (2022) U2Fusion: a unified unsupervised image fusion network. IEEE Trans Pattern Anal Mach Intell 44:502\u2013518. https:\/\/doi.org\/10.1109\/TPAMI.2020.3012548","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"8916_CR36","doi-asserted-by":"crossref","unstructured":"Guo Y, Chen J, Wang J et al (2020) Closed-loop matters: Dual regression networks for single image super-resolution. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5407\u20135416","DOI":"10.1109\/CVPR42600.2020.00545"},{"key":"8916_CR37","doi-asserted-by":"publisher","first-page":"2274","DOI":"10.1109\/TPAMI.2012.120","volume":"34","author":"R Achanta","year":"2012","unstructured":"Achanta R, Shaji A, Smith K et al (2012) SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans Pattern Anal Mach Intell 34:2274\u20132282","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"8916_CR38","doi-asserted-by":"crossref","unstructured":"Lee H, Jeon J, Kim J, Lee S (2017) Structure-texture decomposition of images with interval gradient. In: Computer graphics forum. Wiley Online Library, New York, pp 262\u2013274","DOI":"10.1111\/cgf.12875"},{"key":"8916_CR39","doi-asserted-by":"crossref","unstructured":"Zhai Y, Shah M (2006) Visual attention detection in video sequences using spatiotemporal cues. In: Proceedings of the 14th ACM international conference on Multimedia. pp 815\u2013824","DOI":"10.1145\/1180639.1180824"},{"key":"8916_CR40","doi-asserted-by":"publisher","unstructured":"Toet, Alexander (2014) TNO image fusion dataset. figshare. Dataset. https:\/\/doi.org\/10.6084\/m9.figshare.1008029.v2","DOI":"10.6084\/m9.figshare.1008029.v2"},{"key":"8916_CR41","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li H, Wu X-J (2018) DenseFuse: a fusion approach to infrared and visible images. IEEE Trans Image Process 28:2614\u20132623","journal-title":"IEEE Trans Image Process"},{"key":"8916_CR42","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang Y, Liu Y, Sun P et al (2020) IFCNN: a general image fusion framework based on convolutional neural network. Information Fusion 54:99\u2013118","journal-title":"Information Fusion"},{"key":"8916_CR43","doi-asserted-by":"crossref","unstructured":"Ha Q, Watanabe K, Karasawa T et al (2017) MFNet: towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In: 2017 IEEE\/RSJ international conference on intelligent robots and systems, IROS 2017, Vancouver, BC, Canada, September 24\u201328, 2017. IEEE, New York, pp 5108\u20135115","DOI":"10.1109\/IROS.2017.8206396"},{"key":"8916_CR44","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1049\/el:20020212","volume":"38","author":"G Qu","year":"2002","unstructured":"Qu G, Zhang D, Yan P (2002) Information measure for performance of image fusion. Electron Lett 38:1","journal-title":"Electron Lett"},{"key":"8916_CR45","doi-asserted-by":"publisher","first-page":"2117","DOI":"10.1109\/TIP.2005.859389","volume":"14","author":"HR Sheikh","year":"2005","unstructured":"Sheikh HR, Bovik AC, De Veciana G (2005) An information fidelity criterion for image quality assessment using natural scene statistics. IEEE Trans Image Process 14:2117\u20132128","journal-title":"IEEE Trans Image Process"},{"key":"8916_CR46","doi-asserted-by":"publisher","first-page":"430","DOI":"10.1109\/TIP.2005.859378","volume":"15","author":"HR Sheikh","year":"2006","unstructured":"Sheikh HR, Bovik AC (2006) Image information and visual quality. IEEE Trans Image Process 15:430\u2013444","journal-title":"IEEE Trans Image Process"},{"key":"8916_CR47","doi-asserted-by":"publisher","first-page":"1421","DOI":"10.1016\/j.imavis.2007.12.002","volume":"27","author":"Y Chen","year":"2009","unstructured":"Chen Y, Blum RS (2009) A new automated quality assessment algorithm for image fusion. Image Vis Comput 27:1421\u20131432","journal-title":"Image Vis Comput"},{"key":"8916_CR48","doi-asserted-by":"publisher","first-page":"2959","DOI":"10.1109\/26.477498","volume":"43","author":"AM Eskicioglu","year":"1995","unstructured":"Eskicioglu AM, Fisher PS (1995) Image quality measures and their performance. IEEE Trans Commun 43:2959\u20132965","journal-title":"IEEE Trans Commun"},{"key":"8916_CR49","doi-asserted-by":"crossref","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779\u2013788","DOI":"10.1109\/CVPR.2016.91"},{"key":"8916_CR50","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107404","volume":"106","author":"X Qin","year":"2020","unstructured":"Qin X, Zhang Z, Huang C et al (2020) U2-Net: going deeper with nested U-structure for salient object detection. Pattern Recogn 106:107404","journal-title":"Pattern Recogn"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-023-08916-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-023-08916-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-023-08916-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,9,16]],"date-time":"2023-09-16T15:07:48Z","timestamp":1694876868000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-023-08916-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,13]]},"references-count":50,"journal-issue":{"issue":"30","published-print":{"date-parts":[[2023,10]]}},"alternative-id":["8916"],"URL":"https:\/\/doi.org\/10.1007\/s00521-023-08916-z","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"value":"0941-0643","type":"print"},{"value":"1433-3058","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,8,13]]},"assertion":[{"value":"9 January 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 July 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 August 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}