{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,3]],"date-time":"2025-12-03T03:39:11Z","timestamp":1764733151606,"version":"3.37.3"},"reference-count":33,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T00:00:00Z","timestamp":1733097600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T00:00:00Z","timestamp":1733097600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"The National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62203224"],"award-info":[{"award-number":["62203224"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100019053","name":"Capacity Building Plan for some Non-military Universities and Colleges of Shanghai Scientific Committee","doi-asserted-by":"publisher","award":["22010501300"],"award-info":[{"award-number":["22010501300"]}],"id":[{"id":"10.13039\/501100019053","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,1]]},"DOI":"10.1007\/s11760-024-03631-1","type":"journal-article","created":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T16:19:30Z","timestamp":1733156370000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Infrared and visible image fusion based on saliency detection and deep multi-scale orientational features"],"prefix":"10.1007","volume":"19","author":[{"given":"Gang","family":"Liu","sequence":"first","affiliation":[]},{"given":"Menghan","family":"Jia","sequence":"additional","affiliation":[]},{"given":"Xiao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Durga Prasad","family":"Bavirisetti","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,2]]},"reference":[{"key":"3631_CR1","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"Jiayi Ma","year":"2019","unstructured":"Ma, Jiayi, Ma, Yong, Li, Chang: Infrared and visible image fusion methods and applications: a survey. Inform. Fus. 45, 153\u2013178 (2019)","journal-title":"Inform. Fus."},{"key":"3631_CR2","doi-asserted-by":"publisher","first-page":"477","DOI":"10.1016\/j.infrared.2014.09.019","volume":"67","author":"Jun Wang","year":"2014","unstructured":"Wang, Jun, Peng, Jinye, Feng, Xiaoyi, He, Guiqing, Fan, Jianping: Fusion method for infrared and visible images by using non-negative sparse representation. Infrared Phys. Technol. 67, 477\u2013489 (2014)","journal-title":"Infrared Phys. Technol."},{"key":"3631_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijleo.2022.169599","volume":"266","author":"Yao Peng","year":"2022","unstructured":"Peng, Yao, Liu, Gang, Xiang, Xu., Bavirisetti, Durga Prasad, Xinjie, Gu., Zhang, Xiangbo: Mfdetection: a highly generalized object detection network unified with multilevel heterogeneous image fusion. Optik 266, 169599 (2022)","journal-title":"Optik"},{"issue":"5","key":"3631_CR4","doi-asserted-by":"publisher","first-page":"2531","DOI":"10.1007\/s11760-022-02470-2","volume":"17","author":"G Tirumala Vasu","year":"2023","unstructured":"Tirumala Vasu, G., Palanisamy, P.: Gradient-based multi-focus image fusion using foreground and background pattern recognition with weighted anisotropic diffusion filter. Signal Image Video Process. 17(5), 2531\u20132543 (2023)","journal-title":"Signal Image Video Process."},{"key":"3631_CR5","doi-asserted-by":"crossref","unstructured":"Horng, Shi-Jinn, Zeng, Yun-Jhu: Improving embedded target tracking systems based on siamese networks with infrared images. In 2022 IEEE 4th international conference on artificial intelligence circuits and systems (AICAS), pages 419\u2013422. IEEE (2022)","DOI":"10.1109\/AICAS54282.2022.9869937"},{"issue":"3","key":"3631_CR6","doi-asserted-by":"publisher","first-page":"032203","DOI":"10.1117\/1.JRS.15.032203","volume":"15","author":"Yue Pan","year":"2021","unstructured":"Pan, Yue, Pi, Dechang, Chen, Junfu, Chen, Yang: Remote sensing image fusion with multistream deep rescnn. J. Appl. Remote Sens. 15(3), 032203\u2013032203 (2021)","journal-title":"J. Appl. Remote Sens."},{"key":"3631_CR7","doi-asserted-by":"crossref","unstructured":"Xin, L., Jie, S., Yuehui, W., Ming, L.: Damage detection based on multi-wavelet basis and multi-scale feature fusion. In 2021 international conference on machine learning and intelligent systems engineering (MLISE), pp. 210\u2013213. IEEE (2021)","DOI":"10.1109\/MLISE54096.2021.00044"},{"key":"3631_CR8","first-page":"1","volume":"71","author":"Xiaoling Li","year":"2022","unstructured":"Li, Xiaoling, Chen, Houjin, Li, Yanfeng, Peng, Yahui: Mafusion: multiscale attention network for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 71, 1\u201316 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"4","key":"3631_CR9","doi-asserted-by":"publisher","first-page":"343","DOI":"10.1109\/TMI.2002.1000258","volume":"21","author":"Sabine Dippel","year":"2002","unstructured":"Dippel, Sabine, Stahl, Martin, Wiemker, Rafael, Blaffert, Thomas: Multiscale contrast enhancement for radiographies: Laplacian pyramid versus fast wavelet transform. IEEE Trans. Med. Imag. 21(4), 343\u2013353 (2002)","journal-title":"IEEE Trans. Med. Imag."},{"issue":"18","key":"3631_CR10","doi-asserted-by":"publisher","first-page":"15689","DOI":"10.1007\/s00521-022-07242-0","volume":"34","author":"Shaolei Liu","year":"2022","unstructured":"Liu, Shaolei, Linhao, Qu., Qiao, Qin, Wang, Manning, Song, Zhijian: Wavelet-based self-supervised learning for multi-scene image fusion. Neural Comput. Appl. 34(18), 15689\u201315704 (2022)","journal-title":"Neural Comput. Appl."},{"key":"3631_CR11","doi-asserted-by":"publisher","DOI":"10.1016\/j.sigpro.2021.108062","volume":"184","author":"Xiaosong Li","year":"2021","unstructured":"Li, Xiaosong, Zhou, Fuqiang, Tan, Haishu, Chen, Yuanze, Zuo, Wangxia: Multi-focus image fusion based on nonsubsampled contourlet transform and residual removal. Signal Process. 184, 108062 (2021)","journal-title":"Signal Process."},{"issue":"5","key":"3631_CR12","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"Hui Li","year":"2018","unstructured":"Li, Hui, Xiao-Jun, Wu.: Densefuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"03","key":"3631_CR13","doi-asserted-by":"publisher","first-page":"1850018","DOI":"10.1142\/S0219691318500182","volume":"16","author":"Y Liu","year":"2018","unstructured":"Liu, Y., Chen, X., Juan, C., Peng, H., Zengfu, W.: Infrared and visible image fusion with convolutional neural networks. Int. J. Wavel. Multiresolution Inform. Process. 16(03), 1850018 (2018)","journal-title":"Int. J. Wavel. Multiresolution Inform. Process."},{"key":"3631_CR14","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"Hao Zhang","year":"2021","unstructured":"Zhang, Hao, Le, Zhuliang, Shao, Zhenfeng, Han, Xu., Ma, Jiayi: MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inform. Fusion 66, 40\u201353 (2021)","journal-title":"Inform. Fusion"},{"key":"3631_CR15","doi-asserted-by":"publisher","first-page":"110","DOI":"10.1016\/j.inffus.2021.02.019","volume":"72","author":"Fu Yu","year":"2021","unstructured":"Yu, Fu., Xiao-Jun, Wu., Durrani, Tariq: Image fusion based on generative adversarial network consistent with perception. Inform. Fusion 72, 110\u2013125 (2021)","journal-title":"Inform. Fusion"},{"key":"3631_CR16","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu, Y., Xun, C., Peng, H., Zengfu, W.: Multi-focus image fusion with a deep convolutional neural network. Inform. Fusion 36, 191\u2013207 (2017)","journal-title":"Inform. Fusion"},{"issue":"6","key":"3631_CR17","doi-asserted-by":"publisher","first-page":"1856","DOI":"10.1109\/TMI.2019.2959609","volume":"39","author":"Md Zongwei Zhou","year":"2019","unstructured":"Zongwei Zhou, Md., Siddiquee, Mahfuzur Rahman, Tajbakhsh, Nima, Liang, Jianming: UNet++: Redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imag. 39(6), 1856\u20131867 (2019)","journal-title":"IEEE Trans. Med. Imag."},{"key":"3631_CR18","doi-asserted-by":"crossref","unstructured":"Zhao, J.X., Liu, J.J., Fan, D.P., Cao, Y., Yang, J., Cheng, M.M.: Egnet: Edge guidance network for salient object detection. In Proceedings of the IEEE\/CVF international conference on computer vision, pp. 8779\u20138788 (2019)","DOI":"10.1109\/ICCV.2019.00887"},{"key":"3631_CR19","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1016\/j.infrared.2017.02.005","volume":"82","author":"Jinlei Ma","year":"2017","unstructured":"Ma, Jinlei, Zhou, Zhiqiang, Wang, Bo., Zong, Hua: Infrared and visible image fusion based on visual saliency map and weighted least square optimization. Infrared Phys. Technol. 82, 8\u201317 (2017)","journal-title":"Infrared Phys. Technol."},{"key":"3631_CR20","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1016\/j.infrared.2017.04.018","volume":"83","author":"CH Liu","year":"2017","unstructured":"Liu, C.H., Qi, Y., Ding, W.R.: Infrared and visible image fusion method based on saliency detection in sparse domain. Infrared Phys. Technol. 83, 94\u2013102 (2017)","journal-title":"Infrared Phys. Technol."},{"key":"3631_CR21","first-page":"1","volume":"70","author":"Jiayi Ma","year":"2021","unstructured":"Ma, Jiayi, Tang, Linfeng, Meilong, Xu., Zhang, Hao, Xiao, Guobao: Stdfusionnet: An infrared and visible image fusion network based on salient target detection. IEEE Trans. Instrum. Meas. 70, 1\u201313 (2021)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"3631_CR22","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"Jiayi Ma","year":"2019","unstructured":"Ma, Jiayi, Wei, Yu., Liang, Pengwei, Li, Chang, Jiang, Junjun: Fusiongan: a generative adversarial network for infrared and visible image fusion. Inform. Fusion 48, 11\u201326 (2019)","journal-title":"Inform. Fusion"},{"key":"3631_CR23","doi-asserted-by":"publisher","first-page":"1134","DOI":"10.1109\/TCI.2021.3119954","volume":"7","author":"Hao Zhang","year":"2021","unstructured":"Zhang, Hao, Yuan, Jiteng, Tian, Xin, Ma, Jiayi: GAN-FM: Infrared and visible image fusion using GAN with full-scale skip connection and dual Markovian discriminators. IEEE Trans. Comput. Imag. 7, 1134\u20131147 (2021)","journal-title":"IEEE Trans. Comput. Imag."},{"key":"3631_CR24","doi-asserted-by":"crossref","unstructured":"Du G, Cao X, Liang J, Chen X, Zhan Y.: Medical image segmentation based on u-net: A review. J. Imag. Sci. Technol. 64(2) (2020)","DOI":"10.2352\/J.ImagingSci.Technol.2020.64.2.020508"},{"issue":"12","key":"3631_CR25","doi-asserted-by":"publisher","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","volume":"69","author":"Hui Li","year":"2020","unstructured":"Li, Hui, Xiao-Jun, Wu., Durrani, Tariq: NestFuse: An infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models. IEEE Trans. Instrum. Meas. 69(12), 9645\u20139656 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"3631_CR26","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"Hui Li","year":"2021","unstructured":"Li, Hui, Xiao-Jun, Wu., Kittler, Josef: RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inform. Fusion 73, 72\u201386 (2021)","journal-title":"Inform. Fusion"},{"key":"3631_CR27","doi-asserted-by":"publisher","first-page":"227","DOI":"10.1016\/j.infrared.2017.05.007","volume":"83","author":"Yu Zhang","year":"2017","unstructured":"Zhang, Yu., Zhang, Lijia, Bai, Xiangzhi, Zhang, Li.: Infrared and visual image fusion through infrared feature extraction and visual information preservation. Infrared Phys. Technol. 83, 227\u2013237 (2017)","journal-title":"Infrared Phys. Technol."},{"key":"3631_CR28","doi-asserted-by":"publisher","first-page":"1193","DOI":"10.1007\/s11760-013-0556-9","volume":"9","author":"BK Shreyamsha Kumar","year":"2015","unstructured":"Shreyamsha Kumar, B.K.: Image fusion based on pixel significance using cross bilateral filter. Signal Image Video Process. 9, 1193\u20131204 (2015)","journal-title":"Signal Image Video Process."},{"key":"3631_CR29","doi-asserted-by":"crossref","unstructured":"Bavirisetti, D.P., Xiao, G., Liu, G.: Multi-sensor image fusion based on fourth order partial differential equations. In 2017 20th International conference on information fusion (Fusion), pp. 1\u20139. IEEE (2017)","DOI":"10.23919\/ICIF.2017.8009719"},{"key":"3631_CR30","unstructured":"Li, H., Wu, X.J.: Infrared and visible image fusion using latent low-rank representation. arXiv preprint arXiv:1804.08992 (2018)"},{"key":"3631_CR31","doi-asserted-by":"crossref","unstructured":"Ram\u00a0Prabhakar, K., Sai\u00a0Srikar, V., Venkatesh\u00a0Babu, R.: Deepfuse: A deep unsupervised approach for exposure fusion with extreme exposure image pairs. In Proceedings of the IEEE international conference on computer vision, pp. 4714\u20134722 (2017)","DOI":"10.1109\/ICCV.2017.505"},{"key":"3631_CR32","doi-asserted-by":"crossref","unstructured":"Ma, Jiayi, Zhang, Hao, Shao, Zhenfeng, Liang, Pengwei, Han, Xu.: GANMcC: A generative adversarial network with multiclassification constraints for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 70, 1\u201314 (2020)","DOI":"10.1109\/TIM.2020.3038013"},{"key":"3631_CR33","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101870","volume":"99","author":"Linfeng Tang","year":"2023","unstructured":"Tang, Linfeng, Zhang, Hao, Han, Xu., Ma, Jiayi: Rethinking the necessity of image fusion in high-level vision tasks: a practical infrared and visible image fusion network based on progressive semantic injection and scene fidelity. Inform. Fusion 99, 101870 (2023)","journal-title":"Inform. Fusion"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03631-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-024-03631-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03631-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,28]],"date-time":"2025-01-28T17:52:13Z","timestamp":1738086733000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-024-03631-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,2]]},"references-count":33,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,1]]}},"alternative-id":["3631"],"URL":"https:\/\/doi.org\/10.1007\/s11760-024-03631-1","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2024,12,2]]},"assertion":[{"value":"3 September 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 October 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 October 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 December 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Declaration on ethical approval is not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval"}}],"article-number":"36"}}