{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,25]],"date-time":"2025-11-25T05:09:32Z","timestamp":1764047372150,"version":"3.44.0"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T00:00:00Z","timestamp":1750896000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T00:00:00Z","timestamp":1750896000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s11760-025-04409-9","type":"journal-article","created":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T10:37:05Z","timestamp":1750934225000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["An end-to-end dense network based multi-modal image fusion model for improved object detection in night time images"],"prefix":"10.1007","volume":"19","author":[{"given":"Ankan","family":"Banerjee","sequence":"first","affiliation":[]},{"given":"Dipti","family":"Patra","sequence":"additional","affiliation":[]},{"given":"Pradipta","family":"Roy","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,26]]},"reference":[{"key":"4409_CR1","doi-asserted-by":"crossref","unstructured":"Vajpayee, P., Panigrahy, C. & Kumar, A. Medical image fusion by adaptive gaussian pcnn and improved roberts operator. Signal, Image and Video Processing 1\u20139 (2023)","DOI":"10.1007\/s11760-023-02581-4"},{"key":"4409_CR2","doi-asserted-by":"publisher","first-page":"1127","DOI":"10.3390\/s17051127","volume":"17","author":"Y Zuo","year":"2017","unstructured":"Zuo, Y., Liu, J., Bai, G., Wang, X., Sun, M.: Airborne infrared and visible image fusion combined with region segmentation. Sensors 17, 1127 (2017)","journal-title":"Sensors"},{"key":"4409_CR3","doi-asserted-by":"publisher","first-page":"137","DOI":"10.1016\/j.inffus.2008.08.008","volume":"10","author":"AC Muller","year":"2009","unstructured":"Muller, A.C., Narayanan, S.: Cognitively-engineered multisensor image fusion for military applications. Information Fusion 10, 137\u2013149 (2009)","journal-title":"Information Fusion"},{"key":"4409_CR4","doi-asserted-by":"crossref","unstructured":"Candan, A.\u00a0T. & Kalkan, H. U-net-based rgb and lidar image fusion for road segmentation. Signal, Image and Video Processing 1\u20137 (2023)","DOI":"10.1109\/ASYU56188.2022.9925421"},{"key":"4409_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10586-024-04829-1","volume":"28","author":"L Zhou","year":"2025","unstructured":"Zhou, L., Dong, Y., Ma, B., Yin, Z., Lu, F.: Object detection in low-light conditions based on dbs-yolov8. Clust. Comput. 28, 1\u201320 (2025)","journal-title":"Clust. Comput."},{"key":"4409_CR6","doi-asserted-by":"crossref","unstructured":"Teena, M., Manickavasagan, A.: in Thermal infrared imaging 147\u2013173 (publisherSpringer, 2014)","DOI":"10.1007\/978-3-642-54888-8_8"},{"key":"4409_CR7","doi-asserted-by":"crossref","unstructured":"Lin, C., Jiang, Z., Cong, J., Zou, L.: Rnn with high precision and noise immunity: A robust and learning-free method for beamforming. IEEE Internet of Things Journal (2025)","DOI":"10.1109\/JIOT.2025.3529532"},{"key":"4409_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11227-024-06569-w","volume":"81","author":"H Xiao","year":"2025","unstructured":"Xiao, H., Chen, X., Luo, L., Lin, C.: A dual-path feature reuse multi-scale network for remote sensing image super-resolution. J. Supercomput. 81, 1\u201328 (2025)","journal-title":"J. Supercomput."},{"key":"4409_CR9","doi-asserted-by":"publisher","first-page":"4425","DOI":"10.1007\/s11831-021-09540-7","volume":"28","author":"H Kaur","year":"2021","unstructured":"Kaur, H., Koundal, D., Kadyan, V.: Image fusion techniques: a survey. Archives of computational methods in Engineering 28, 4425\u20134447 (2021)","journal-title":"Archives of computational methods in Engineering"},{"key":"4409_CR10","doi-asserted-by":"publisher","first-page":"514","DOI":"10.1109\/JSTSP.2008.2001309","volume":"2","author":"C-Y Chen","year":"2008","unstructured":"Chen, C.-Y., Lin, T.-M., Wolf, W.H.: A visible\/infrared fusion algorithm for distributed smart cameras. IEEE Journal of Selected Topics in Signal Processing 2, 514\u2013525 (2008)","journal-title":"IEEE Journal of Selected Topics in Signal Processing"},{"key":"4409_CR11","doi-asserted-by":"crossref","unstructured":"Liu, J.: et\u00a0al.Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection, 5802\u20135811 (2022)","DOI":"10.1109\/CVPR52688.2022.00571"},{"key":"4409_CR12","doi-asserted-by":"crossref","unstructured":"Goyal, B., Dogra, A., Khoond, R., Gupta, A., Anand, R.: Infrared and visible image fusion for concealed weapon detection using transform and spatial domain filters, 1\u20134 (IEEE, 2021)","DOI":"10.1109\/ICRITO51393.2021.9596074"},{"key":"4409_CR13","doi-asserted-by":"publisher","first-page":"237","DOI":"10.26599\/JICV.2023.9210018","volume":"6","author":"Q Yuan","year":"2023","unstructured":"Yuan, Q., et al.: Enhanced target tracking algorithm for autonomous driving based on visible and infrared image fusion. Journal of Intelligent and Connected Vehicles 6, 237\u2013249 (2023)","journal-title":"Journal of Intelligent and Connected Vehicles"},{"key":"4409_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2025.110024","volume":"232","author":"X Liu","year":"2025","unstructured":"Liu, X., Lv, M., Ma, C., Fu, Z., Zhang, L.: Multi-modal image fusion of visible and infrared for precise positioning of uavs in agricultural fields. Comput. Electron. Agric. 232, 110024 (2025)","journal-title":"Comput. Electron. Agric."},{"key":"4409_CR15","doi-asserted-by":"publisher","first-page":"10385","DOI":"10.1038\/s41598-023-37295-7","volume":"13","author":"BS Krishnan","year":"2023","unstructured":"Krishnan, B.S., et al.: Fusion of visible and thermal images improves automated detection and classification of animals for drone surveys. Sci. Rep. 13, 10385 (2023)","journal-title":"Sci. Rep."},{"key":"4409_CR16","doi-asserted-by":"publisher","first-page":"432","DOI":"10.1109\/LGRS.2012.2207944","volume":"10","author":"J Marcello","year":"2012","unstructured":"Marcello, J., Medina, A., Eugenio, F.: Evaluation of spatial and spectral effectiveness of pixel-level fusion techniques. IEEE Geosci. Remote Sens. Lett. 10, 432\u2013436 (2012)","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"4409_CR17","doi-asserted-by":"publisher","first-page":"24829","DOI":"10.1109\/JSEN.2021.3113579","volume":"21","author":"Y Yang","year":"2021","unstructured":"Yang, Y., et al.: Multi-sensor fusion of infrared and visible images based on modified side window filter and intensity transformation. IEEE Sens. J. 21, 24829\u201324843 (2021)","journal-title":"IEEE Sens. J."},{"key":"4409_CR18","doi-asserted-by":"publisher","first-page":"162","DOI":"10.1109\/TCI.2022.3151472","volume":"8","author":"H Yan","year":"2022","unstructured":"Yan, H., Zhang, J.-X., Zhang, X.: Injected infrared and visible image fusion via $$ l_ \\{1\\}$$ decomposition model and guided filtering. IEEE Transactions on Computational Imaging 8, 162\u2013173 (2022)","journal-title":"IEEE Transactions on Computational Imaging"},{"key":"4409_CR19","doi-asserted-by":"crossref","unstructured":"Jian, B.-L., Tu, C.-C.: Multi-level optimal fusion algorithm for infrared and visible image. Signal, Image and Video Processing 1\u20139 (2023)","DOI":"10.1007\/s11760-023-02653-5"},{"key":"4409_CR20","doi-asserted-by":"crossref","unstructured":"Ram\u00a0Prabhakar, K., Sai Srikar, V., Venkatesh Babu, R.: Deepfuse: A deep unsupervised approach for exposure fusion with extreme exposure image pairs, 4714\u20134722 (2017)","DOI":"10.1109\/ICCV.2017.505"},{"key":"4409_CR21","doi-asserted-by":"crossref","unstructured":"Li, H., Wu, X.-J., Kittler, J.: Infrared and visible image fusion using a deep learning framework, 2705\u20132710 (IEEE, 2018)","DOI":"10.1109\/ICPR.2018.8546006"},{"key":"4409_CR22","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.-J.: Densefuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 28, 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"4409_CR23","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: Fusiongan: A generative adversarial network for infrared and visible image fusion. Information fusion 48, 11\u201326 (2019)","journal-title":"Information fusion"},{"key":"4409_CR24","doi-asserted-by":"publisher","first-page":"640","DOI":"10.1109\/TCI.2020.2965304","volume":"6","author":"R Hou","year":"2020","unstructured":"Hou, R., et al.: Vif-net: an unsupervised framework for infrared and visible image fusion. IEEE Transactions on Computational Imaging 6, 640\u2013651 (2020)","journal-title":"IEEE Transactions on Computational Imaging"},{"key":"4409_CR25","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44, 502\u2013518 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4409_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijleo.2021.168084","volume":"248","author":"J Zhou","year":"2021","unstructured":"Zhou, J., et al.: An infrared and visible image fusion method based on vgg-19 network. Optik 248, 168084 (2021)","journal-title":"Optik"},{"key":"4409_CR27","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Wu, X.-J., Kittler, J.: Rfn-nest: An end-to-end residual fusion network for infrared and visible images. Information Fusion 73, 72\u201386 (2021)","journal-title":"Information Fusion"},{"key":"4409_CR28","doi-asserted-by":"crossref","unstructured":"Wang, Z., Shao, W., Chen, Y., Xu, J., Zhang, X.: Infrared and visible image fusion via interactive compensatory attention adversarial learning. IEEE Transactions on Multimedia (2022)","DOI":"10.1109\/TMM.2022.3228685"},{"key":"4409_CR29","doi-asserted-by":"crossref","unstructured":"Zhao, W., Xie, S., Zhao, F., He, Y., Lu, H.: Metafusion: Infrared and visible image fusion via meta-feature embedding from object detection, 13955\u201313965 (2023)","DOI":"10.1109\/CVPR52729.2023.01341"},{"key":"4409_CR30","doi-asserted-by":"crossref","unstructured":"Gu, X.: et\u00a0al. Infrared-visible synthetic data from game engine for image fusion improvement. IEEE Transactions on Games (2023)","DOI":"10.1109\/TG.2023.3263001"},{"key":"4409_CR31","doi-asserted-by":"crossref","unstructured":"Zhao, Z.: et\u00a0al.Equivariant multi-modality image fusion, 25912\u201325921 (2024)","DOI":"10.1109\/CVPR52733.2024.02448"},{"key":"4409_CR32","doi-asserted-by":"publisher","first-page":"723","DOI":"10.1016\/j.procs.2024.04.069","volume":"235","author":"SPK Reddy","year":"2024","unstructured":"Reddy, S.P.K., Harikiran, J., Chandana, B.S.: Deep cnn based multi object detection and tracking in video frames with mean distributed feature set. Procedia Computer Science 235, 723\u2013734 (2024)","journal-title":"Procedia Computer Science"},{"key":"4409_CR33","doi-asserted-by":"publisher","first-page":"3212","DOI":"10.1109\/TNNLS.2018.2876865","volume":"30","author":"Z-Q Zhao","year":"2019","unstructured":"Zhao, Z.-Q., Zheng, P., Xu, S.-T., Wu, X.: Object detection with deep learning: A review. IEEE transactions on neural networks and learning systems 30, 3212\u20133232 (2019)","journal-title":"IEEE transactions on neural networks and learning systems"},{"key":"4409_CR34","doi-asserted-by":"crossref","unstructured":"Dutta, UK. Seeing objects in dark with continual contrastive learning, 286\u2013302 (Springer, 2023)","DOI":"10.1007\/978-3-031-25075-0_21"},{"key":"4409_CR35","doi-asserted-by":"publisher","first-page":"1134","DOI":"10.1109\/TCI.2021.3119954","volume":"7","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Yuan, J., Tian, X., Ma, J.: Gan-fm: Infrared and visible image fusion using gan with full-scale skip connection and dual markovian discriminators. IEEE Transactions on Computational Imaging 7, 1134\u20131147 (2021)","journal-title":"IEEE Transactions on Computational Imaging"},{"key":"4409_CR36","doi-asserted-by":"publisher","DOI":"10.1016\/j.optlastec.2022.108466","volume":"156","author":"G Li","year":"2022","unstructured":"Li, G., Lai, W., Qu, X.: Pedestrian detection based on light perception fusion of visible and thermal images. Optics & Laser Technology 156, 108466 (2022)","journal-title":"Optics & Laser Technology"},{"key":"4409_CR37","doi-asserted-by":"crossref","unstructured":"Xue, S., Liu, Y., Xu, C., Li, J. Object detection in visible and infrared missile borne fusion image, 19\u201323 (IEEE, 2022)","DOI":"10.1109\/ICICML57342.2022.10009652"},{"key":"4409_CR38","unstructured":"Toet, A.: TNO Image Fusion Dataset (2022). https:\/\/figshare.com\/articles\/dataset\/TNO_Image_Fusion_Dataset\/1008029"},{"key":"4409_CR39","doi-asserted-by":"crossref","unstructured":"Jia, X., Zhu, C., Li, M., Tang, W., Zhou, W.: Llvip: A visible-infrared paired dataset for low-light vision, 3496\u20133504 (2021)","DOI":"10.1109\/ICCVW54120.2021.00389"},{"key":"4409_CR40","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1016\/j.inffus.2016.02.001","volume":"31","author":"J Ma","year":"2016","unstructured":"Ma, J., Chen, C., Li, C., Huang, J.: Infrared and visible image fusion via gradient transfer and total variation minimization. Information Fusion 31, 100\u2013109 (2016)","journal-title":"Information Fusion"},{"key":"4409_CR41","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L.: & Weinberger, K.\u00a0Q. Densely connected convolutional networks, 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"4409_CR42","first-page":"1","volume":"20","author":"M Wang","year":"2023","unstructured":"Wang, M., He, W., Zhang, H.: A spatial-spectral transformer network with total variation loss for hyperspectral image denoising. IEEE Geosci. Remote Sens. Lett. 20, 1\u20135 (2023)","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"4409_CR43","doi-asserted-by":"crossref","unstructured":"Tan, M., Pang, R., Le, QV.: Efficientdet: Scalable and efficient object detection, 10781\u201310790 (2020)","DOI":"10.1109\/CVPR42600.2020.01079"},{"key":"4409_CR44","unstructured":"Wu, Y., Kirillov, A., Massa, F., Lo, W.-Y., Girshick, R.: Detectron2 (2019)"},{"key":"4409_CR45","first-page":"107984","volume":"37","author":"A Wang","year":"2024","unstructured":"Wang, A., et al.: Yolov10: Real-time end-to-end object detection. Adv. Neural. Inf. Process. Syst. 37, 107984\u2013108011 (2024)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"4409_CR46","doi-asserted-by":"crossref","unstructured":"Carion, N.: et\u00a0al.End-to-end object detection with transformers, 213\u2013229 (Springer, 2020)","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"4409_CR47","doi-asserted-by":"crossref","unstructured":"Lin, TY.: et\u00a0al.Microsoft coco: Common objects in context, 740\u2013755 (Springer, 2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"4409_CR48","doi-asserted-by":"crossref","unstructured":"Pandey, B., Joshi, D., Arora, A.S., Upadhyay, N., Chhabra, H.: A deep learning approach for automated detection and segmentation of pressure ulcers using infrared-based thermal imaging. IEEE Sens. J. 22, 14762\u201314768 (2022)","DOI":"10.1109\/JSEN.2022.3184105"},{"key":"4409_CR49","doi-asserted-by":"crossref","unstructured":"Shreyamsha Kumar, B.: Image fusion based on pixel significance using cross bilateral filter. SIViP 9, 1193\u20131204 (2015)","DOI":"10.1007\/s11760-013-0556-9"},{"key":"4409_CR50","doi-asserted-by":"crossref","unstructured":"Zhang, X., Ye, P., Xiao, G.: Vifb: A visible and infrared image fusion benchmark, 104\u2013105 (2020)","DOI":"10.1016\/j.infrared.2019.103124"},{"key":"4409_CR51","doi-asserted-by":"publisher","first-page":"193","DOI":"10.1016\/j.inffus.2005.10.001","volume":"8","author":"H Chen","year":"2007","unstructured":"Chen, H., Varshney, P.K.: A human perception inspired quality metric for image fusion based on regional information. Information fusion 8, 193\u2013207 (2007)","journal-title":"Information fusion"},{"key":"4409_CR52","doi-asserted-by":"publisher","first-page":"133","DOI":"10.1016\/j.aqpro.2015.02.019","volume":"4","author":"P Jagalingam","year":"2015","unstructured":"Jagalingam, P., Hegde, A.V.: A review of quality metrics for fused image. Aquatic Procedia 4, 133\u2013142 (2015)","journal-title":"Aquatic Procedia"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04409-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-04409-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04409-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T22:51:35Z","timestamp":1757199095000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-04409-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,26]]},"references-count":52,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["4409"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-04409-9","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2025,6,26]]},"assertion":[{"value":"5 September 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 June 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 June 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 June 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval"}},{"value":"Not applicable","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"Not applicable","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}],"article-number":"808"}}