{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T05:39:49Z","timestamp":1774589989343,"version":"3.50.1"},"reference-count":60,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2025,1,20]],"date-time":"2025-01-20T00:00:00Z","timestamp":1737331200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,20]],"date-time":"2025-01-20T00:00:00Z","timestamp":1737331200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Double First-Class Innovation Research Project for the People\u2019s Public Security University of China","award":["2023SYL08"],"award-info":[{"award-number":["2023SYL08"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s00371-024-03792-7","type":"journal-article","created":{"date-parts":[[2025,1,20]],"date-time":"2025-01-20T03:53:39Z","timestamp":1737345219000},"page":"7063-7078","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["Bridging visible and infrared modalities: a dual-level joint align network for person re-identification"],"prefix":"10.1007","volume":"41","author":[{"given":"Mengnan","family":"Hu","sequence":"first","affiliation":[]},{"given":"Qianli","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Rong","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,20]]},"reference":[{"key":"3792_CR1","doi-asserted-by":"publisher","unstructured":"Wu, A., Zheng, W.S., Yu, H.X., et al.: RGB-infrared cross-modality person re-identification. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 5390\u20135399 (2017). https:\/\/doi.org\/10.1109\/iccv.2017.575","DOI":"10.1109\/iccv.2017.575"},{"issue":"4","key":"3792_CR2","doi-asserted-by":"publisher","first-page":"1958","DOI":"10.1109\/TNNLS.2021.3105702","volume":"34","author":"H Liu","year":"2023","unstructured":"Liu, H., Ma, S., Xia, D., et al.: SFANet: a spectrum-aware feature augmentation network for visible-infrared person reidentification. IEEE Trans. Neural Netw. Learn. Syst. 34(4), 1958\u20131971 (2023). https:\/\/doi.org\/10.1109\/TNNLS.2021.3105702","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"3792_CR3","doi-asserted-by":"publisher","unstructured":"Liu, J., Sun, Y., Zhu, F., et al.: Learning memory-augmented unidirectional metrics for cross-modality person re-identification. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 19366\u201319375 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.01876","DOI":"10.1109\/CVPR52688.2022.01876"},{"issue":"2","key":"3792_CR4","doi-asserted-by":"publisher","first-page":"1631","DOI":"10.1609\/aaai.v37i2.25250","volume":"37","author":"Y Ling","year":"2023","unstructured":"Ling, Y., Zhong, Z., Luo, Z., et al.: Cross-modality earth mover\u2019s distance for visible thermal person re-identification. Proc. AAAI Conf. Artif. Intell. 37(2), 1631\u20131639 (2023). https:\/\/doi.org\/10.1609\/aaai.v37i2.25250","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"issue":"6","key":"3792_CR5","doi-asserted-by":"publisher","first-page":"5104","DOI":"10.1109\/TCSVT.2023.3338813","volume":"34","author":"Y Ling","year":"2023","unstructured":"Ling, Y., Zhong, Z., Luo, Z., et al.: Bridge gap in pixel and feature level for cross-modality person re-identification. IEEE Trans. Circuits Syst. Video Technol. 34(6), 5104\u20135117 (2023). https:\/\/doi.org\/10.1109\/TCSVT.2023.3338813","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"4","key":"3792_CR6","doi-asserted-by":"publisher","first-page":"2363","DOI":"10.1007\/s00371-023-02923-w","volume":"40","author":"Y Wang","year":"2024","unstructured":"Wang, Y., Xu, K., Chai, Y., et al.: Semantic consistent feature construction and multi-granularity feature learning for visible-infrared person re-identification. Vis. Comput. 40(4), 2363\u20132379 (2024). https:\/\/doi.org\/10.1007\/s00371-023-02923-w","journal-title":"Vis. Comput."},{"key":"3792_CR7","doi-asserted-by":"publisher","unstructured":"Dai, H., Xie, Q., Ma, Y., et al.: RGB-infrared person re-identification via image modality conversion. In: 2020 25th International Conference on Pattern Recognition (ICPR), pp. 592\u2013598. IEEE (2021). https:\/\/doi.org\/10.1109\/ICPR48806.2021.9412257","DOI":"10.1109\/ICPR48806.2021.9412257"},{"key":"3792_CR8","doi-asserted-by":"publisher","unstructured":"Wang, G., Zhang, T., Cheng, J., et al.: RGB-infrared cross-modality person re-identification via joint pixel and feature alignment. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 3622\u20133631 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00372","DOI":"10.1109\/ICCV.2019.00372"},{"key":"3792_CR9","doi-asserted-by":"publisher","unstructured":"Wang, Z., Wang, Z., Zheng, Y., et al.: Learning to reduce dual-level discrepancy for infrared-visible person re-identification. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 618\u2013626 (2019). https:\/\/doi.org\/10.1109\/CVPR.2019.00071","DOI":"10.1109\/CVPR.2019.00071"},{"key":"3792_CR10","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.neucom.2021.02.088","volume":"443","author":"D Xia","year":"2021","unstructured":"Xia, D., Liu, H., Xu, L., et al.: Visible-infrared person re-identification with data augmentation via cycle-consistent adversarial network. Neurocomputing 443, 35\u201346 (2021). https:\/\/doi.org\/10.1016\/j.neucom.2021.02.088","journal-title":"Neurocomputing"},{"key":"3792_CR11","doi-asserted-by":"publisher","unstructured":"Kim, M., Kim, S., Park, J., et al.: PartMix: regularization strategy to learn part discovery for visible-infrared person re-identification. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 18621\u201318632 (2023). https:\/\/doi.org\/10.1109\/CVPR52729.2023.01786","DOI":"10.1109\/CVPR52729.2023.01786"},{"key":"3792_CR12","doi-asserted-by":"publisher","unstructured":"Ye, M., Ruan, W., Du, B., et al.: Channel augmented joint learning for visible-infrared recognition. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 13567\u201313576 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.01331","DOI":"10.1109\/ICCV48922.2021.01331"},{"key":"3792_CR13","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2024.3377252","author":"T Liang","year":"2024","unstructured":"Liang, T., Jin, Y., Liu, W., et al.: Bridging the gap: multi-level cross-modality joint alignment for visible-infrared person re-identification. IEEE Trans. Circuits Syst. Video Technol. (2024). https:\/\/doi.org\/10.1109\/TCSVT.2024.3377252","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"3792_CR14","doi-asserted-by":"publisher","unstructured":"Tan, L., Zhang, Y., Shen, S., et al.: Exploring invariant representation for visible-infrared person re-identification. abs\/2302.00884 (2023). https:\/\/doi.org\/10.48550\/arXiv.2302.00884","DOI":"10.48550\/arXiv.2302.00884"},{"key":"3792_CR15","doi-asserted-by":"publisher","unstructured":"Zhang, Y., Wang, H.: Diverse embedding expansion network and low-light cross-modality benchmark for visible-infrared person re-identification. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2153\u20132162 (2023). https:\/\/doi.org\/10.1109\/CVPR52729.2023.00214","DOI":"10.1109\/CVPR52729.2023.00214"},{"key":"3792_CR16","doi-asserted-by":"publisher","unstructured":"Cheng, K., Hua, X., Lu, H., et al.: Multi-scale semantic correlation mining for visible-infrared person re-identification. abs\/2311.14395 (2023). https:\/\/doi.org\/10.48550\/arXiv.2311.14395","DOI":"10.48550\/arXiv.2311.14395"},{"key":"3792_CR17","doi-asserted-by":"publisher","unstructured":"Ren, K., Zhang, L.J.A.: Implicit discriminative knowledge learning for visible-infrared person re-identification. abs\/2403.11708 (2024). https:\/\/doi.org\/10.48550\/arXiv.2403.11708","DOI":"10.48550\/arXiv.2403.11708"},{"key":"3792_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2024.3504240","volume":"2024","author":"R Zhang","year":"2024","unstructured":"Zhang, R., Liu, G., Zhang, Q., et al.: Detail-aware network for infrared image enhancement. IEEE Trans. Geosci. Remote Sens. 2024, 1\u20131 (2024). https:\/\/doi.org\/10.1109\/TGRS.2024.3504240","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"3792_CR19","doi-asserted-by":"publisher","unstructured":"Zhang, Q., Lai, C., Liu, J., et al.: FMCNet: feature-level modality compensation for visible-infrared person re-identification. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7339\u20137348 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.00720","DOI":"10.1109\/CVPR52688.2022.00720"},{"key":"3792_CR20","doi-asserted-by":"publisher","unstructured":"Sun, H., Liu, J., Zhang, Z., et al.: Not all pixels are matched: dense contrastive learning for cross-modality person re-identification. In: Proceedings of the 30th ACM international conference on multimedia, pp. 5333\u20135341 (2022). https:\/\/doi.org\/10.1145\/3503161.3547970","DOI":"10.1145\/3503161.3547970"},{"key":"3792_CR21","doi-asserted-by":"publisher","first-page":"346","DOI":"10.1109\/LSP.2024.3354190","volume":"31","author":"Y Qian","year":"2024","unstructured":"Qian, Y., Tang, S.K.: Pose attention-guided paired-images generation for visible-infrared person re-identification. IEEE Signal Process. Lett. 31, 346\u2013350 (2024)","journal-title":"IEEE Signal Process. Lett."},{"key":"3792_CR22","doi-asserted-by":"publisher","unstructured":"Li, D., Wei, X., Hong, X., et al.: Infrared-visible cross-modal person re-identification with an x modality. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 4610\u20134617 (2020). https:\/\/doi.org\/10.1609\/aaai.v34i04.5891","DOI":"10.1609\/aaai.v34i04.5891"},{"issue":"3","key":"3792_CR23","doi-asserted-by":"publisher","first-page":"1418","DOI":"10.1109\/TCSVT.2021.3072171","volume":"32","author":"X Zhong","year":"2022","unstructured":"Zhong, X., Lu, T., Huang, W., et al.: Grayscale enhancement colorization network for visible-infrared person re-identification. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1418\u20131430 (2022). https:\/\/doi.org\/10.1109\/TCSVT.2021.3072171","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"3792_CR24","doi-asserted-by":"publisher","unstructured":"Wei, Z., Yang, X., Wang, N., et al. Syncretic modality collaborative learning for visible infrared person re-identification. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 225\u2013234 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.00029","DOI":"10.1109\/ICCV48922.2021.00029"},{"key":"3792_CR25","doi-asserted-by":"publisher","unstructured":"Zhang, Y., Yan, Y., Lu, Y., et al. Towards a unified middle modality learning for visible-infrared person re-identification. In: Proceedings of the 29th ACM international conference on multimedia, pp. 788\u2013796 (2021). https:\/\/doi.org\/10.1145\/3474085.3475250","DOI":"10.1145\/3474085.3475250"},{"key":"3792_CR26","doi-asserted-by":"publisher","first-page":"10","DOI":"10.1016\/j.patcog.2024.110873","volume":"157","author":"Z Qian","year":"2025","unstructured":"Qian, Z., Lin, Y., Du, B.: Visible\u2013infrared person re-identification via patch-mixed cross-modality learning. Pattern Recogn. 157, 10 (2025). https:\/\/doi.org\/10.1016\/j.patcog.2024.110873","journal-title":"Pattern Recogn."},{"key":"3792_CR27","doi-asserted-by":"publisher","unstructured":"Lee, D., Lee, J., Lee, J., et al. Regularization strategy for point cloud via rigidly mixed sample. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 15895\u201315904 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.01564","DOI":"10.1109\/CVPR46437.2021.01564"},{"key":"3792_CR28","doi-asserted-by":"publisher","unstructured":"Shen, Z., Liu, Z., Liu, Z., et al.: Un-mix: rethinking image mixtures for unsupervised visual representation learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 2216\u20132224 (2022). https:\/\/doi.org\/10.1609\/aaai.v36i2.20119","DOI":"10.1609\/aaai.v36i2.20119"},{"key":"3792_CR29","doi-asserted-by":"publisher","unstructured":"Kim, J.-H., Choo, W., Song, H.O.: Puzzle mix: exploiting saliency and local statistics for optimal mixup. In: Proceedings of the 37th International Conference on Machine Learning, pp. 5275\u20135285 (2020). https:\/\/doi.org\/10.48550\/arXiv.2009.06962","DOI":"10.48550\/arXiv.2009.06962"},{"issue":"3","key":"3792_CR30","doi-asserted-by":"publisher","first-page":"e2248","DOI":"10.1002\/cav.2248","volume":"35","author":"W Zhao","year":"2024","unstructured":"Zhao, W., Zhu, J., Huang, J., et al.: GAN-based multi-decomposition photo cartoonization. Comput. Anim. Virtual Worlds 35(3), e2248 (2024). https:\/\/doi.org\/10.1002\/cav.2248","journal-title":"Comput. Anim. Virtual Worlds"},{"key":"3792_CR31","doi-asserted-by":"publisher","unstructured":"Kingma, D.P., Welling, M.J.C.: Auto-encoding variational Bayes. abs\/1312.6114 (2013). https:\/\/doi.org\/10.48550\/arXiv.1312.6114","DOI":"10.48550\/arXiv.1312.6114"},{"key":"3792_CR32","doi-asserted-by":"publisher","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., et al.: High-resolution image synthesis with latent diffusion models. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10674\u201310685 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.01042","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"3792_CR33","doi-asserted-by":"publisher","unstructured":"Zhang, L., Rao, A., Agrawala, M.J.I.C.I.C.O.C.V.: Adding conditional control to text-to-image diffusion models. In: 2023 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 3836\u20133847 (2023). https:\/\/doi.org\/10.1109\/ICCV51070.2023.00355","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"3792_CR34","doi-asserted-by":"publisher","unstructured":"Wu, W., Zhao, Y., Chen, H., et al.: DatasetDM: synthesizing data with perception annotations using diffusion models. abs\/2308.06160 (2023). https:\/\/doi.org\/10.48550\/arXiv.2308.06160","DOI":"10.48550\/arXiv.2308.06160"},{"key":"3792_CR35","doi-asserted-by":"publisher","unstructured":"Huang, H., Huang, Y., Wang, L.J.A.: VI-Diff: unpaired visible-infrared translation diffusion model for single modality labeled visible-infrared person re-identification. abs\/2310.04122 (2023). https:\/\/doi.org\/10.48550\/arXiv.2310.04122","DOI":"10.48550\/arXiv.2310.04122"},{"key":"3792_CR36","doi-asserted-by":"publisher","unstructured":"Parmar, G., Park, T., Narasimhan, S., et al. One-step image translation with text-to-image models. abs\/2403.12036 (2024). https:\/\/doi.org\/10.48550\/arXiv.2403.12036","DOI":"10.48550\/arXiv.2403.12036"},{"issue":"6","key":"3792_CR37","doi-asserted-by":"publisher","first-page":"509","DOI":"10.1016\/j.vrih.2023.06.003","volume":"5","author":"M Wang","year":"2023","unstructured":"Wang, M., Meng, M., Liu, J., et al.: Learning adequate alignment and interaction for cross-modal retrieval. Virtual Real. Intell. Hardw. 5(6), 509\u2013522 (2023). https:\/\/doi.org\/10.1016\/j.vrih.2023.06.003","journal-title":"Virtual Real. Intell. Hardw."},{"key":"3792_CR38","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2021","unstructured":"Lin, X., Sun, S., Huang, W., et al.: EAPT: efficient attention pyramid transformer for image processing. IEEE Trans. Multimed. 25, 50\u201361 (2021). https:\/\/doi.org\/10.1109\/TMM.2021.3120873","journal-title":"IEEE Trans. Multimed."},{"key":"3792_CR39","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2110.13985","author":"A Gu","year":"2021","unstructured":"Gu, A., Johnson, I., Goel, K., et al.: Combining recurrent, convolutional, and continuous-time models with linear state-space layers. Neural Inf. Process. Syst. (2021). https:\/\/doi.org\/10.48550\/arXiv.2110.13985","journal-title":"Neural Inf. Process. Syst."},{"key":"3792_CR40","doi-asserted-by":"publisher","unstructured":"Gu, A., Goel, K., R'e, C.J.A.: Efficiently modeling long sequences with structured state spaces. abs\/2111.00396 (2021). https:\/\/doi.org\/10.48550\/arXiv.2111.00396","DOI":"10.48550\/arXiv.2111.00396"},{"key":"3792_CR41","doi-asserted-by":"publisher","unstructured":"Smith, J., Warrington, A., Linderman, S.W.J.A.: Simplified state space layers for sequence modeling. abs\/2208.04933 (2022). https:\/\/doi.org\/10.48550\/arXiv.2208.04933","DOI":"10.48550\/arXiv.2208.04933"},{"key":"3792_CR42","doi-asserted-by":"publisher","unstructured":"Gu, A., Dao, T.J.A.: Mamba: linear-time sequence modeling with selective state spaces. abs\/2312.00752 (2023). https:\/\/doi.org\/10.48550\/arXiv.2312.00752","DOI":"10.48550\/arXiv.2312.00752"},{"key":"3792_CR43","doi-asserted-by":"publisher","unstructured":"Zhu, L., Liao, B., Zhang, Q., et al.: Vision mamba: efficient visual representation learning with bidirectional state space model. abs\/2401.09417 (2024). https:\/\/doi.org\/10.48550\/arXiv.2401.09417","DOI":"10.48550\/arXiv.2401.09417"},{"key":"3792_CR44","doi-asserted-by":"publisher","unstructured":"Liu, Y., Tian, Y., Zhao, Y., et al.: VMamba: visual state space model. abs\/2401.10166 (2024). https:\/\/doi.org\/10.48550\/arXiv.2401.10166","DOI":"10.48550\/arXiv.2401.10166"},{"key":"3792_CR45","doi-asserted-by":"publisher","unstructured":"Pei, X., Huang, T., Xu, C.J.A.: EfficientVMamba: atrous selective scan for light weight visual mamba. abs\/2403.09977 (2024). https:\/\/doi.org\/10.48550\/arXiv.2403.09977","DOI":"10.48550\/arXiv.2403.09977"},{"key":"3792_CR46","doi-asserted-by":"publisher","unstructured":"Wang, Z., Li, C., Xu, H., et al.: Mamba YOLO: SSMs-based YOLO for object detection. abs\/2406.05835 (2024). https:\/\/doi.org\/10.48550\/arXiv.2406.05835","DOI":"10.48550\/arXiv.2406.05835"},{"key":"3792_CR47","doi-asserted-by":"publisher","unstructured":"Shaker, A., Wasim, S.T., Khan, S., et al.: GroupMamba: parameter-efficient and accurate group visual state space model. abs\/2407.13772 (2024). https:\/\/doi.org\/10.48550\/arXiv.2407.13772","DOI":"10.48550\/arXiv.2407.13772"},{"key":"3792_CR48","doi-asserted-by":"publisher","unstructured":"Wang, J., Zhu, W., Wang, P., et al.: Selective structured state-spaces for long-form video understanding. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6387\u20136397 (2023). https:\/\/doi.org\/10.1109\/CVPR52729.2023.00618","DOI":"10.1109\/CVPR52729.2023.00618"},{"key":"3792_CR49","doi-asserted-by":"publisher","unstructured":"Peng, S., Zhu, X., Deng, H., et al. FusionMamba: efficient image fusion with state space model. abs\/2404.07932 (2024). https:\/\/doi.org\/10.48550\/arXiv.2404.07932","DOI":"10.48550\/arXiv.2404.07932"},{"key":"3792_CR50","doi-asserted-by":"publisher","unstructured":"Hu, E.J., Shen, Y., Wallis, P., et al.: Lora: low-rank adaptation of large language models. abs\/2106.09685 (2021). https:\/\/doi.org\/10.48550\/arXiv.2106.09685","DOI":"10.48550\/arXiv.2106.09685"},{"key":"3792_CR51","doi-asserted-by":"publisher","unstructured":"Zhang, R., Isola, P., Efros, A.A., et al.: The unreasonable effectiveness of deep features as a perceptual metric. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00068","DOI":"10.1109\/CVPR.2018.00068"},{"key":"3792_CR52","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016). https:\/\/doi.org\/10.48550\/arXiv.1512.03385","DOI":"10.48550\/arXiv.1512.03385"},{"key":"3792_CR53","doi-asserted-by":"publisher","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., et al. Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 6000\u20136010 (2017). https:\/\/doi.org\/10.48550\/arXiv.1706.03762","DOI":"10.48550\/arXiv.1706.03762"},{"issue":"6","key":"3792_CR54","doi-asserted-by":"publisher","first-page":"2872","DOI":"10.1109\/TPAMI.2021.3054775","volume":"44","author":"M Ye","year":"2022","unstructured":"Ye, M., Shen, J., Lin, G., et al.: Deep learning for person re-identification: a survey and outlook. IEEE Trans. Pattern Anal. Mach. Intell. 44(6), 2872\u20132893 (2022). https:\/\/doi.org\/10.1109\/TPAMI.2021.3054775","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"3792_CR55","doi-asserted-by":"publisher","first-page":"605","DOI":"10.3390\/s17030605","volume":"17","author":"DT Nguyen","year":"2017","unstructured":"Nguyen, D.T., Hong, H.G., Kim, K.W., et al.: Person recognition system based on a combination of body images from visible light and thermal cameras. Sensors 17(3), 605 (2017). https:\/\/doi.org\/10.3390\/s17030605","journal-title":"Sensors"},{"key":"3792_CR56","doi-asserted-by":"publisher","unstructured":"Deng, J., Dong, W., Socher, R., et al.: ImageNet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255 (2009). https:\/\/doi.org\/10.1109\/CVPR.2009.5206848","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"3792_CR57","doi-asserted-by":"publisher","unstructured":"Zhong, Z., Zheng, L., Kang, G., et al.: Random erasing data augmentation. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 13001\u201313008 (2020). https:\/\/doi.org\/10.1609\/aaai.v34i07.7000","DOI":"10.1609\/aaai.v34i07.7000"},{"key":"3792_CR58","doi-asserted-by":"publisher","unstructured":"Ye, M., Lan, X., Li, J., et al.: Hierarchical discriminative learning for visible thermal person re-identification. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32(1) (2018). https:\/\/doi.org\/10.1609\/aaai.v32i1.12293","DOI":"10.1609\/aaai.v32i1.12293"},{"key":"3792_CR59","doi-asserted-by":"publisher","unstructured":"Ye, M., Shen, J.J., Crandall, D., et al.: Dynamic dual-attentive aggregation learning for visible-infrared person re-identification. In: Computer Vision\u2014ECCV 2020, pp. 229\u2013247. Springer, Cham. https:\/\/doi.org\/10.1007\/978-3-030-58520-4_14","DOI":"10.1007\/978-3-030-58520-4_14"},{"issue":"11","key":"3792_CR60","first-page":"2579","volume":"9","author":"LVD Maaten","year":"2008","unstructured":"Maaten, L.V.D., Hinton, G.E.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9(11), 2579\u20132605 (2008)","journal-title":"J. Mach. Learn. Res."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03792-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-024-03792-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03792-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T04:13:50Z","timestamp":1757132030000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-024-03792-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,20]]},"references-count":60,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["3792"],"URL":"https:\/\/doi.org\/10.1007\/s00371-024-03792-7","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1,20]]},"assertion":[{"value":"30 December 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 January 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}