{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T15:34:45Z","timestamp":1772120085617,"version":"3.50.1"},"reference-count":46,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T00:00:00Z","timestamp":1750896000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T00:00:00Z","timestamp":1750896000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["Grant 62302142 and Grant 61876057"],"award-info":[{"award-number":["Grant 62302142 and Grant 61876057"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["Grant 62302142 and Grant 61876057"],"award-info":[{"award-number":["Grant 62302142 and Grant 61876057"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["Grant 62302142 and Grant 61876057"],"award-info":[{"award-number":["Grant 62302142 and Grant 61876057"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["Grant 62302142 and Grant 61876057"],"award-info":[{"award-number":["Grant 62302142 and Grant 61876057"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["Grant 62302142 and Grant 61876057"],"award-info":[{"award-number":["Grant 62302142 and Grant 61876057"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002858","name":"China Postdoctoral Science Foundation","doi-asserted-by":"publisher","award":["Grant 2022M720981"],"award-info":[{"award-number":["Grant 2022M720981"]}],"id":[{"id":"10.13039\/501100002858","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002858","name":"China Postdoctoral Science Foundation","doi-asserted-by":"publisher","award":["Grant 2022M720981"],"award-info":[{"award-number":["Grant 2022M720981"]}],"id":[{"id":"10.13039\/501100002858","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002858","name":"China Postdoctoral Science Foundation","doi-asserted-by":"publisher","award":["Grant 2022M720981"],"award-info":[{"award-number":["Grant 2022M720981"]}],"id":[{"id":"10.13039\/501100002858","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002858","name":"China Postdoctoral Science Foundation","doi-asserted-by":"publisher","award":["Grant 2022M720981"],"award-info":[{"award-number":["Grant 2022M720981"]}],"id":[{"id":"10.13039\/501100002858","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002858","name":"China Postdoctoral Science Foundation","doi-asserted-by":"publisher","award":["Grant 2022M720981"],"award-info":[{"award-number":["Grant 2022M720981"]}],"id":[{"id":"10.13039\/501100002858","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003995","name":"Natural Science Foundation of Anhui Province","doi-asserted-by":"publisher","award":["No.2208085MF158"],"award-info":[{"award-number":["No.2208085MF158"]}],"id":[{"id":"10.13039\/501100003995","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003995","name":"Natural Science Foundation of Anhui Province","doi-asserted-by":"publisher","award":["No.2208085MF158"],"award-info":[{"award-number":["No.2208085MF158"]}],"id":[{"id":"10.13039\/501100003995","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003995","name":"Natural Science Foundation of Anhui Province","doi-asserted-by":"publisher","award":["No.2208085MF158"],"award-info":[{"award-number":["No.2208085MF158"]}],"id":[{"id":"10.13039\/501100003995","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003995","name":"Natural Science Foundation of Anhui Province","doi-asserted-by":"publisher","award":["No.2208085MF158"],"award-info":[{"award-number":["No.2208085MF158"]}],"id":[{"id":"10.13039\/501100003995","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003995","name":"Natural Science Foundation of Anhui Province","doi-asserted-by":"publisher","award":["No.2208085MF158"],"award-info":[{"award-number":["No.2208085MF158"]}],"id":[{"id":"10.13039\/501100003995","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Key Research Plan of Anhui Province - Strengthening Police with Science and Technology","award":["Grant 202004d07020012"],"award-info":[{"award-number":["Grant 202004d07020012"]}]},{"name":"Key Research Plan of Anhui Province - Strengthening Police with Science and Technology","award":["Grant 202004d07020012"],"award-info":[{"award-number":["Grant 202004d07020012"]}]},{"name":"Key Research Plan of Anhui Province - Strengthening Police with Science and Technology","award":["Grant 202004d07020012"],"award-info":[{"award-number":["Grant 202004d07020012"]}]},{"name":"Key Research Plan of Anhui Province - Strengthening Police with Science and Technology","award":["Grant 202004d07020012"],"award-info":[{"award-number":["Grant 202004d07020012"]}]},{"name":"Key Research Plan of Anhui Province - Strengthening Police with Science and Technology","award":["Grant 202004d07020012"],"award-info":[{"award-number":["Grant 202004d07020012"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. Mach. Learn. &amp; Cyber."],"published-print":{"date-parts":[[2025,11]]},"DOI":"10.1007\/s13042-025-02717-7","type":"journal-article","created":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T02:04:44Z","timestamp":1750903484000},"page":"8893-8906","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Robust visible-infrared person re-identification via frequency-space joint disentanglement and fusion network"],"prefix":"10.1007","volume":"16","author":[{"given":"Rui","family":"Sun","sequence":"first","affiliation":[]},{"given":"Xuebin","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Guoxi","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Long","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Libing","family":"Qian","sequence":"additional","affiliation":[]},{"given":"Jun","family":"Gao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,26]]},"reference":[{"key":"2717_CR1","first-page":"3520","volume":"35","author":"Z Zhao","year":"2021","unstructured":"Zhao Z, Liu B, Chu Q et al (2021) Joint color-irrelevant consistency learning and identity-aware modality adaptation for visible-infrared cross modality person re-identification. Proc AAAI Conf Artif Intell 35:3520\u20133528","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"2717_CR2","first-page":"1835","volume":"37","author":"H Lu","year":"2023","unstructured":"Lu H, Zou X, Zhang P (2023) Learning progressive modality-shared transformers for effective visible-infrared person re-identification. Proc AAAI Conf Artif Intell 37:1835\u20131843","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"2717_CR3","doi-asserted-by":"crossref","unstructured":"Chen Y, Wan L, Li Z, et\u00a0al (2021) Neural feature search for rgb-infrared person re-identification. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 587\u2013597","DOI":"10.1109\/CVPR46437.2021.00065"},{"key":"2717_CR4","doi-asserted-by":"publisher","first-page":"381","DOI":"10.1007\/978-3-031-19809-0_22","volume-title":"Computer Vision - ECCV 2022","author":"X Li","year":"2022","unstructured":"Li X, Lu Y, Liu B et al (2022) Counterfactual intervention feature transfer for visible-infrared person re-identification. In: Avidan S, Brostow G, Ciss\u00e9 M et al (eds) Computer Vision - ECCV 2022. Springer Nature Switzerland, Cham, pp 381\u2013398"},{"key":"2717_CR5","doi-asserted-by":"publisher","first-page":"4414","DOI":"10.1109\/TMM.2020.3042080","volume":"23","author":"H Liu","year":"2021","unstructured":"Liu H, Tan X, Zhou X (2021) Parameter sharing exploration and hetero-center triplet loss for visible-thermal person re-identification. IEEE Trans Multimed 23:4414\u20134425","journal-title":"IEEE Trans Multimed"},{"key":"2717_CR6","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1016\/j.neucom.2019.12.100","volume":"386","author":"Y Zhu","year":"2020","unstructured":"Zhu Y, Yang Z, Wang L et al (2020) Hetero-center loss for cross-modality person re-identification. Neurocomputing 386:97\u2013109","journal-title":"Neurocomputing"},{"key":"2717_CR7","doi-asserted-by":"publisher","first-page":"480","DOI":"10.1007\/978-3-031-19781-9_28","volume-title":"Computer Vision - ECCV 2022","author":"K Jiang","year":"2022","unstructured":"Jiang K, Zhang T, Liu X et al (2022) Cross-modality transformer for visible-infrared person re-identification. In: Avidan S, Brostow G, Ciss\u00e9 M et al (eds) Computer Vision - ECCV 2022. Springer Nature Switzerland, Cham, pp 480\u2013496"},{"key":"2717_CR8","doi-asserted-by":"publisher","first-page":"728","DOI":"10.1109\/TIFS.2020.3001665","volume":"16","author":"M Ye","year":"2021","unstructured":"Ye M, Shen J, Shao L (2021) Visible-infrared person re-identification via homogeneous augmented tri-modal learning. IEEE Trans Inf Forensics Secur 16:728\u2013739","journal-title":"IEEE Trans Inf Forensics Secur"},{"key":"2717_CR9","doi-asserted-by":"crossref","unstructured":"Yu H, Cheng X, Peng W, et\u00a0al (2023) Modality unifying network for visible-infrared person re-identification. 2023 IEEE\/CVF International Conference on Computer Vision (ICCV) pp 11151\u201311161","DOI":"10.1109\/ICCV51070.2023.01027"},{"key":"2717_CR10","unstructured":"Goodfellow IJ, Pouget-Abadie J, Mirza M, et\u00a0al (2014) Generative adversarial nets. In: Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2. MIT Press, Cambridge, MA, USA, NIPS\u201914, p 2672\u20132680"},{"key":"2717_CR11","doi-asserted-by":"publisher","first-page":"357","DOI":"10.1049\/ep.1965.0268","volume":"11","author":"K Cattermole","year":"1965","unstructured":"Cattermole K (1965) The fourier transform and its applications. Electron Power 11:357\u2013357","journal-title":"Electron Power"},{"key":"2717_CR12","unstructured":"Khayam SA (2003) The discrete cosine transform (dct): Theory and application1. Course Notes, Department of Electrical & Computer Engineering"},{"key":"2717_CR13","doi-asserted-by":"crossref","unstructured":"Lee S, Bae J, Kim HY (2023) Decompose, adjust, compose: Effective normalization by playing with frequency for domain generalization. 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) pp 11776\u201311785","DOI":"10.1109\/CVPR52729.2023.01133"},{"key":"2717_CR14","doi-asserted-by":"publisher","first-page":"181","DOI":"10.1007\/978-3-031-19800-7_11","volume-title":"Computer Vision - ECCV 2022","author":"H Yu","year":"2022","unstructured":"Yu H, Zheng N, Zhou M et al (2022) Frequency and spatial dual guidance for image dehazing. In: Avidan S, Brostow G, Ciss\u00e9 M et al (eds) Computer Vision - ECCV 2022. Springer Nature Switzerland, Cham, pp 181\u2013198"},{"key":"2717_CR15","doi-asserted-by":"publisher","first-page":"1655","DOI":"10.1109\/TPAMI.2018.2846566","volume":"41","author":"F Radenovic","year":"2017","unstructured":"Radenovic F, Tolias G, Chum O (2017) Fine-tuning cnn image retrieval with no human annotation. IEEE Trans Pattern Anal Mach Intell 41:1655\u20131668","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"6","key":"2717_CR16","doi-asserted-by":"publisher","first-page":"2872","DOI":"10.1109\/TPAMI.2021.3054775","volume":"44","author":"M Ye","year":"2022","unstructured":"Ye M, Shen J, Lin G et al (2022) Deep learning for person re-identification: A survey and outlook. IEEE Trans Pattern Anal Mach Intell 44(6):2872\u20132893","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"2717_CR17","doi-asserted-by":"crossref","unstructured":"Chen D, Xu D, Li H, et\u00a0al (2018) Group consistent similarity learning via deep crf for person re-identification. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 8649\u20138658","DOI":"10.1109\/CVPR.2018.00902"},{"issue":"9","key":"2717_CR18","doi-asserted-by":"publisher","first-page":"3391","DOI":"10.1109\/TCSVT.2020.3043026","volume":"31","author":"X Ning","year":"2021","unstructured":"Ning X, Gong K, Li W et al (2021) Feature refinement and filter network for person re-identification. IEEE Trans Circuits Syst Video Technol 31(9):3391\u20133402","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"2717_CR19","doi-asserted-by":"crossref","unstructured":"Luo H, Gu Y, Liao X, et\u00a0al (2019) Bag of tricks and a strong baseline for deep person re-identification. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp 1487\u20131495","DOI":"10.1109\/CVPRW.2019.00190"},{"key":"2717_CR20","doi-asserted-by":"crossref","unstructured":"Liao S, Shao L (2022) Graph sampling based deep metric learning for generalizable person re-identification. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 7349\u20137358","DOI":"10.1109\/CVPR52688.2022.00721"},{"key":"2717_CR21","doi-asserted-by":"crossref","unstructured":"Zeng K, Ning M, Wang Y, et\u00a0al (2020) Hierarchical clustering with hard-batch triplet loss for person re-identification. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 13654\u201313662","DOI":"10.1109\/CVPR42600.2020.01367"},{"key":"2717_CR22","doi-asserted-by":"crossref","unstructured":"Ye M, Wang Z, Lan X, et\u00a0al (2018) Visible thermal person re-identification via dual-constrained top-ranking. In: Proceedings of the 27th International Joint Conference on Artificial Intelligence. AAAI Press, IJCAI\u201918, p 1092-1099","DOI":"10.24963\/ijcai.2018\/152"},{"key":"2717_CR23","doi-asserted-by":"crossref","unstructured":"Ye M, Shen J, J. Crandall D, et al (2020) Dynamic dual-attentive aggregation learning for visible-infrared person re-identification. In: Vedaldi A, Bischof H, Brox T et al (eds) Computer Vision - ECCV 2020. Springer International Publishing, Cham, pp 229\u2013247","DOI":"10.1007\/978-3-030-58520-4_14"},{"key":"2717_CR24","doi-asserted-by":"crossref","unstructured":"Wu A, Zheng WS, Yu HX, et\u00a0al (2017) Rgb-infrared cross-modality person re-identification. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp 5390\u20135399","DOI":"10.1109\/ICCV.2017.575"},{"key":"2717_CR25","doi-asserted-by":"crossref","unstructured":"Kim M, Kim S, Park J, et\u00a0al (2023) Partmix: Regularization strategy to learn part discovery for visible-infrared person re-identification. 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) pp 18621\u201318632","DOI":"10.1109\/CVPR52729.2023.01786"},{"key":"2717_CR26","doi-asserted-by":"crossref","unstructured":"Zhang Y, Wang H (2023) Diverse embedding expansion network and low-light cross-modality benchmark for visible-infrared person re-identification. 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) pp 2153\u20132162","DOI":"10.1109\/CVPR52729.2023.00214"},{"key":"2717_CR27","unstructured":"Cui Y, Tao Y, Bing Z, et\u00a0al (2023a) Selective frequency network for image restoration. In: International Conference on Learning Representations"},{"key":"2717_CR28","doi-asserted-by":"publisher","first-page":"1093","DOI":"10.1109\/TPAMI.2023.3330416","volume":"46","author":"Y Cui","year":"2023","unstructured":"Cui Y, Ren W, Cao X et al (2023) Image restoration via frequency selection. IEEE Trans Pattern Anal Mach Intell 46:1093\u20131108","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"2717_CR29","doi-asserted-by":"crossref","unstructured":"Cui Y, Ren W, Cao X, et\u00a0al (2023c) Focal network for image restoration. 2023 IEEE\/CVF International Conference on Computer Vision (ICCV) pp 12955\u201312965","DOI":"10.1109\/ICCV51070.2023.01195"},{"key":"2717_CR30","doi-asserted-by":"crossref","unstructured":"Jiang H, Maharjan P, Li Z, et\u00a0al (2022) Dct-based residual network for nir image colorization. In: 2022 IEEE International Conference on Image Processing (ICIP), pp 2926\u20132930","DOI":"10.1109\/ICIP46576.2022.9897373"},{"key":"2717_CR31","unstructured":"Li Y, Zhang T, Zhang Y (2024) Frequency domain modality-invariant feature learning for visible-infrared person re-identification. ArXiv abs\/2401.01839"},{"key":"2717_CR32","doi-asserted-by":"crossref","unstructured":"Cui, Y, Ren, W, Knoll, A (2024). Omni-Kernel Network for Image Restoration. Proceedings of the AAAI Conference on Artificial Intelligence, 38(2), 1426\u20131434","DOI":"10.1609\/aaai.v38i2.27907"},{"key":"2717_CR33","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, et\u00a0al (2016) Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"2717_CR34","unstructured":"Hermans A, Beyer L, Leibe B (2017) In defense of the triplet loss for person re-identification. CoRR abs\/1703.07737"},{"key":"2717_CR35","doi-asserted-by":"crossref","unstructured":"Nguyen TD, Hong HG, Kim KW, et\u00a0al (2017) Person recognition system based on a combination of body images from visible light and thermal cameras. Sensors (Basel, Switzerland) 17","DOI":"10.3390\/s17030605"},{"key":"2717_CR36","doi-asserted-by":"publisher","first-page":"4610","DOI":"10.1609\/aaai.v34i04.5891","volume":"34","author":"D Li","year":"2020","unstructured":"Li D, Wei X, Hong X et al (2020) Infrared-visible cross-modal person re-identification with an x modality. Proceedings of the AAAI Conference on Artificial Intelligence 34:4610\u20134617","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"key":"2717_CR37","doi-asserted-by":"crossref","unstructured":"Chen Y, Wan L, Li Z, et\u00a0al (2021) Neural feature search for rgb-infrared person re-identification. 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) pp 587\u2013597","DOI":"10.1109\/CVPR46437.2021.00065"},{"key":"2717_CR38","doi-asserted-by":"crossref","unstructured":"Ye M, Ruan W, Du B, et\u00a0al (2021) Channel augmented joint learning for visible-infrared recognition. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp 13547\u201313556","DOI":"10.1109\/ICCV48922.2021.01331"},{"key":"2717_CR39","doi-asserted-by":"publisher","first-page":"3668","DOI":"10.1109\/TMM.2022.3163847","volume":"25","author":"J Zhao","year":"2023","unstructured":"Zhao J, Wang H, Zhou Y et al (2023) Spatial-channel enhanced transformer for visible-infrared person re-identification. IEEE Trans Multimedia 25:3668\u20133680","journal-title":"IEEE Trans Multimedia"},{"key":"2717_CR40","doi-asserted-by":"publisher","first-page":"2352","DOI":"10.1109\/TIP.2022.3141868","volume":"31","author":"C Chen","year":"2022","unstructured":"Chen C, Ye M, Qi M et al (2022) Structure-aware positional transformer for visible-infrared person re-identification. IEEE Trans Image Process 31:2352\u20132364","journal-title":"IEEE Trans Image Process"},{"key":"2717_CR41","doi-asserted-by":"crossref","unstructured":"Zhang Q, Lai CJ, Liu J, et\u00a0al (2022) Fmcnet: Feature-level modality compensation for visible-infrared person re-identification. 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) pp 7339\u20137348","DOI":"10.1109\/CVPR52688.2022.00720"},{"issue":"7","key":"2717_CR42","doi-asserted-by":"publisher","first-page":"2511","DOI":"10.1007\/s11263-024-01997-w","volume":"132","author":"M Yang","year":"2024","unstructured":"Yang M, Huang Z, Peng X et al (2024) Robust object re-identification with coupled noisy labels. Int J Comput Vis 132(7):2511\u20132529. https:\/\/doi.org\/10.1007\/s11263-024-01997-w","journal-title":"Int J Comput Vis"},{"key":"2717_CR43","doi-asserted-by":"crossref","unstructured":"Yang M, Huang Z, Hu P, et\u00a0al (2022) Learning with twin noisy labels for visible-infrared person re-identification. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 14288\u201314297","DOI":"10.1109\/CVPR52688.2022.01391"},{"key":"2717_CR44","unstructured":"Chen M, Wang Z, Zheng F (2021) Benchmarks for corruption invariant person re-identification. ArXiv abs\/2111.00880"},{"key":"2717_CR45","doi-asserted-by":"crossref","unstructured":"Wei Z, Yang X, Wang N, et\u00a0al (2021) Syncretic modality collaborative learning for visible infrared person re-identification. 2021 IEEE\/CVF International Conference on Computer Vision (ICCV) pp 225\u2013234","DOI":"10.1109\/ICCV48922.2021.00029"},{"issue":"86","key":"2717_CR46","first-page":"2579","volume":"9","author":"L van der Maaten","year":"2008","unstructured":"van der Maaten L, Hinton G (2008) Visualizing data using t-sne. J Mach Learn Res 9(86):2579\u20132605","journal-title":"J Mach Learn Res"}],"container-title":["International Journal of Machine Learning and Cybernetics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-025-02717-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13042-025-02717-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-025-02717-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,7]],"date-time":"2025-11-07T11:23:35Z","timestamp":1762514615000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13042-025-02717-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,26]]},"references-count":46,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2025,11]]}},"alternative-id":["2717"],"URL":"https:\/\/doi.org\/10.1007\/s13042-025-02717-7","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-5143263\/v1","asserted-by":"object"}]},"ISSN":["1868-8071","1868-808X"],"issn-type":[{"value":"1868-8071","type":"print"},{"value":"1868-808X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,6,26]]},"assertion":[{"value":"24 September 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 June 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 June 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The Author confirms that it is not under consideration for publication elsewhere.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for Publication"}}]}}