{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,21]],"date-time":"2026-02-21T06:06:12Z","timestamp":1771653972486,"version":"3.50.1"},"reference-count":29,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T00:00:00Z","timestamp":1769040000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T00:00:00Z","timestamp":1769040000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["No.U24A2092"],"award-info":[{"award-number":["No.U24A2092"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Qin Chuangyuan \u2019Scientists + Engineers\u2019 Team Construction Program in Shaanxi Province","award":["No.2022KXJ-38"],"award-info":[{"award-number":["No.2022KXJ-38"]}]},{"DOI":"10.13039\/501100013101","name":"Scientific Research Plan Projects of Shaanxi Education Department","doi-asserted-by":"publisher","award":["No. 20JK0758"],"award-info":[{"award-number":["No. 20JK0758"]}],"id":[{"id":"10.13039\/501100013101","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s11760-026-05102-1","type":"journal-article","created":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T00:51:27Z","timestamp":1769043087000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Underground personnel abnormal behavior recognition method based on infrared and visible image fusion"],"prefix":"10.1007","volume":"20","author":[{"given":"Qiang","family":"Guo","sequence":"first","affiliation":[]},{"given":"Junming","family":"Bai","sequence":"additional","affiliation":[]},{"given":"Hongguang","family":"Pan","sequence":"additional","affiliation":[]},{"given":"Huipeng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Ze","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Libin","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Li","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,22]]},"reference":[{"issue":"6","key":"5102_CR1","doi-asserted-by":"publisher","first-page":"9280","DOI":"10.1109\/JIOT.2019.2911669","volume":"6","author":"Z Gao","year":"2019","unstructured":"Gao, Z., Xuan, H.-Z., Zhang, H., Wan, S., Choo, K.-K.R.: Adaptive fusion and category-level dictionary learning model for multiview human action recognition. IEEE Internet Things J. 6(6), 9280\u20139293 (2019)","journal-title":"IEEE Internet Things J."},{"issue":"03","key":"5102_CR2","doi-asserted-by":"publisher","first-page":"1850018","DOI":"10.1142\/S0219691318500182","volume":"16","author":"Y Liu","year":"2018","unstructured":"Liu, Y., Chen, X., Cheng, J., Peng, H., Wang, Z.: Infrared and visible image fusion with convolutional neural networks, International Journal of Wavelets. Multiresolution and Information Processing 16(03), 1850018 (2018)","journal-title":"Multiresolution and Information Processing"},{"key":"5102_CR3","doi-asserted-by":"publisher","first-page":"640","DOI":"10.1109\/TCI.2020.2965304","volume":"6","author":"R Hou","year":"2020","unstructured":"Hou, R., Zhou, D., Nie, R., Liu, D., Xiong, L., Guo, Y., Yu, C.: Vif-net: An unsupervised framework for infrared and visible image fusion. IEEE Transactions on Computational Imaging 6, 640\u2013651 (2020)","journal-title":"IEEE Transactions on Computational Imaging"},{"key":"5102_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.infrared.2022.104405","volume":"127","author":"S Yi","year":"2022","unstructured":"Yi, S., Jiang, G., Liu, X., Li, J., Chen, L.: Tcpmfnet: An infrared and visible image fusion network with composite auto encoder and transformer convolutional parallel mixed fusion strategy. Infrared Physics & Technology 127, 104405 (2022)","journal-title":"Infrared Physics & Technology"},{"key":"5102_CR5","doi-asserted-by":"crossref","unstructured":"Fu, Y., Wu, X.-J.: A dual-branch network for infrared and visible image fusion, in: 2020 25th International Conference on Pattern Recognition (ICPR), 2021, pp. 10675\u201310680","DOI":"10.1109\/ICPR48806.2021.9412293"},{"key":"5102_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.optlaseng.2024.108700","volume":"185","author":"X Zhang","year":"2025","unstructured":"Zhang, X., Liu, G., Xing, M., Wang, G., Bavirisetti, D.P.: Illumination enhancement discriminator and compensation attention based low-light visible and infrared image fusion. Opt. Lasers Eng. 185, 108700 (2025)","journal-title":"Opt. Lasers Eng."},{"key":"5102_CR7","doi-asserted-by":"publisher","first-page":"1350","DOI":"10.1016\/j.infrared.2024.105270","volume":"138","author":"R Chang","year":"2024","unstructured":"Chang, R., Zhao, S., Rao, Y., Yang, Y.: Lvif-net: Learning synchronous visible and infrared image fusion and enhancement under low-light conditions. Infrared Physics & Technology 138, 1350\u20134495 (2024)","journal-title":"Infrared Physics & Technology"},{"key":"5102_CR8","doi-asserted-by":"publisher","first-page":"1568","DOI":"10.1016\/j.asoc.2024.112114","volume":"165","author":"L Sun","year":"2024","unstructured":"Sun, L., Li, Y., Muhammad, G.: Soft computing-driven infrared and visible image fusion network for security application service. Appl. Soft Comput. 165, 1568\u20134946 (2024)","journal-title":"Appl. Soft Comput."},{"key":"5102_CR9","doi-asserted-by":"crossref","unstructured":"Zhang, X., Xu, C., Fan, G., Hua, Z., Li, J., Zhou, J.: Fscmf: A dual-branch frequency-spatial joint perception cross-modality network for visible and infrared image fusion, Neurocomputing (2025) 130376","DOI":"10.1016\/j.neucom.2025.130376"},{"key":"5102_CR10","doi-asserted-by":"crossref","unstructured":"Cui, Y., Duan, P., Li, J.: Dfgic-net: diffusion feature-guided information complementary network for infrared and visible light fusion: Y. cui et al., The Journal of Supercomputing 81\u00a0(8) (2025) 1034","DOI":"10.1007\/s11227-025-07476-4"},{"key":"5102_CR11","first-page":"3639","volume":"2017","author":"R Hinami","year":"2017","unstructured":"Hinami, R., Mei, T., Satoh, S.: Joint detection and recounting of abnormal events by learning deep generic knowledge, in. IEEE International Conference on Computer Vision (ICCV) 2017, 3639\u20133647 (2017)","journal-title":"IEEE International Conference on Computer Vision (ICCV)"},{"key":"5102_CR12","doi-asserted-by":"publisher","first-page":"0957","DOI":"10.1016\/j.eswa.2017.09.029","volume":"91","author":"A Ben Mabrouk","year":"2018","unstructured":"Ben Mabrouk, A., Zagrouba, E.: Abnormal behavior recognition for intelligent video surveillance systems: A review. Expert Syst. Appl. 91, 0957\u20134174 (2018)","journal-title":"Expert Syst. Appl."},{"key":"5102_CR13","doi-asserted-by":"publisher","first-page":"0262","DOI":"10.1016\/j.imavis.2021.104120","volume":"108","author":"K Zhou","year":"2021","unstructured":"Zhou, K., Hui, B., Wang, J., Wang, C., Wu, T.: A study on attention-based lstm for abnormal behavior recognition with variable pooling. Image Vis. Comput. 108, 0262\u20138856 (2021)","journal-title":"Image Vis. Comput."},{"key":"5102_CR14","doi-asserted-by":"crossref","unstructured":"Vidya, M., Selvakumar, Q.\u00a0M. S.:\u00a0 An effective framework of human abnormal behaviour recognition and tracking using multiscale dilated assisted residual attention network, Expert Systems with Applications 247 (2024) 0957\u20134174","DOI":"10.1016\/j.eswa.2024.123264"},{"issue":"3","key":"5102_CR15","doi-asserted-by":"publisher","first-page":"451","DOI":"10.1109\/83.557356","volume":"6","author":"D Jobson","year":"1997","unstructured":"Jobson, D., Rahman, Z., Woodell, G.: Properties and performance of a center\/surround retinex. IEEE Trans. Image Process. 6(3), 451\u2013462 (1997)","journal-title":"IEEE Trans. Image Process."},{"key":"5102_CR16","doi-asserted-by":"crossref","unstructured":"Hao, S., Han, X., Guo, Y., Xu, X., Wang, M.: Low-light image enhancement with semi-decoupled decomposition. IEEE Trans. Multimedia 22(12), 3025\u20133038 (2020)","DOI":"10.1109\/TMM.2020.2969790"},{"key":"5102_CR17","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2024.110490","volume":"153","author":"R Khan","year":"2024","unstructured":"Khan, R., Mehmood, A., Shahid, F., Zheng, Z., Ibrahim, M.M.: Lit me up: A reference free adaptive low light image enhancement for in-the-wild conditions. Pattern Recogn. 153, 110490 (2024)","journal-title":"Pattern Recogn."},{"key":"5102_CR18","unstructured":"Yu, F., Koltun, V.: Multi-scale context aggregation by dilated convolutions (2016). arXiv:1511.07122"},{"key":"5102_CR19","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zou, Y., Shi, W.: Dilated convolution neural network with leakyrelu for environmental sound classification, in: 2017 22nd International Conference on Digital Signal Processing (DSP), 2017, pp. 1\u20135","DOI":"10.1109\/ICDSP.2017.8096153"},{"key":"5102_CR20","doi-asserted-by":"publisher","first-page":"82832","DOI":"10.1109\/ACCESS.2020.2991439","volume":"8","author":"J Qu","year":"2020","unstructured":"Qu, J., Su, C., Zhang, Z., Razi, A.: Dilated convolution and feature fusion ssd network for small object detection in remote sensing images. IEEE Access 8, 82832\u201382843 (2020)","journal-title":"IEEE Access"},{"key":"5102_CR21","doi-asserted-by":"crossref","unstructured":"Chen, J., Hong, H., Song, B., Guo, J., Chen, C., Xu, J.: Mdct: Multi-kernel dilated convolution and transformer for one-stage object detection of remote sensing images, Remote Sensing 15\u00a0(2)","DOI":"10.3390\/rs15020371"},{"key":"5102_CR22","first-page":"4260","volume":"2017","author":"R Zhao","year":"2017","unstructured":"Zhao, R., Ali, H., van der Smagt, P.: Two-stream rnn\/cnn for action recognition in 3d videos, in. IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS) 2017, 4260\u20134267 (2017)","journal-title":"IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)"},{"issue":"12","key":"5102_CR23","doi-asserted-by":"publisher","first-page":"3007","DOI":"10.1109\/TPAMI.2017.2771306","volume":"40","author":"J Liu","year":"2018","unstructured":"Liu, J., Shahroudy, A., Xu, D., Kot, A.C., Wang, G.: Skeleton-based action recognition using spatio-temporal lstm network with trust gates. IEEE Trans. Pattern Anal. Mach. Intell. 40(12), 3007\u20133021 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"5102_CR24","doi-asserted-by":"crossref","unstructured":"Liu, Y. ,Zhang, S., Li, Z., Zhang, Y.: Abnormal behavior recognition based on key points of human skeleton, IFAC-PapersOnLine 53\u00a0(5) (2020) 441\u2013445, 3rd IFAC Workshop on Cyber-Physical & Human Systems CPHS 2020","DOI":"10.1016\/j.ifacol.2021.04.120"},{"key":"5102_CR25","doi-asserted-by":"crossref","unstructured":"Jing, Y., Wang, F.: Tp-vit: A two-pathway vision transformer for video action recognition, in: ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022, pp. 2185\u20132189","DOI":"10.1109\/ICASSP43922.2022.9747276"},{"key":"5102_CR26","doi-asserted-by":"crossref","unstructured":"Niu, X., Zheng, G.: Design of abnormal behavior detection system based on multimodal fusion, Procedia Computer Science 247 (2024) 770\u2013779, the 11th International Conference on Applications and Techniques in Cyber Intelligence","DOI":"10.1016\/j.procs.2024.10.093"},{"key":"5102_CR27","doi-asserted-by":"crossref","unstructured":"Zhu, X., Zhu, Y., Wang, H., Wen, H., Yan, Y., Liu, P.: Skeleton sequence and rgb frame based multi-modality feature fusion network for action recognition, ACM Trans. Multimedia Comput. Commun. Appl. 18\u00a0(3)","DOI":"10.1145\/3491228"},{"issue":"1","key":"5102_CR28","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1049\/cit2.12014","volume":"6","author":"Y Xing","year":"2021","unstructured":"Xing, Y., Zhu, J.: Deep learning-based action recognition with 3d skeleton: A survey. CAAI Transactions on Intelligence Technology 6(1), 80\u201392 (2021)","journal-title":"CAAI Transactions on Intelligence Technology"},{"issue":"17","key":"5102_CR29","doi-asserted-by":"publisher","first-page":"19157","DOI":"10.1109\/JSEN.2021.3089705","volume":"21","author":"X Weiyao","year":"2021","unstructured":"Weiyao, X., Muqing, W., Min, Z., Ting, X.: Fusion of skeleton and rgb features for rgb-d human action recognition. IEEE Sens. J. 21(17), 19157\u201319164 (2021)","journal-title":"IEEE Sens. J."}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-026-05102-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-026-05102-1","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-026-05102-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,21]],"date-time":"2026-02-21T05:26:15Z","timestamp":1771651575000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-026-05102-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,22]]},"references-count":29,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["5102"],"URL":"https:\/\/doi.org\/10.1007\/s11760-026-05102-1","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,22]]},"assertion":[{"value":"30 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 December 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 January 2026","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 January 2026","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there are no personal relationships with other people or organizations that can inappropriately influence our work and no professional or other personal interest of any nature or kind in any service and company that could be construed as influencing.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}},{"value":"This paper does not address any data ethics issues. The research adhered to ethical guidelines and principles. The data used in this study are sourced from publicly available, open-access datasets. All data used in research comply with provider terms, and no identifiable information was used.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Approval"}},{"value":"All authors are fully informed of the study.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed Consent"}}],"article-number":"56"}}