{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T15:25:20Z","timestamp":1773933920724,"version":"3.50.1"},"reference-count":48,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,11,21]],"date-time":"2025-11-21T00:00:00Z","timestamp":1763683200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,21]],"date-time":"2025-11-21T00:00:00Z","timestamp":1763683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100002582","name":"Gwangju Institute of Science and Technology","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100002582","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Cluster Comput"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s10586-025-05792-1","type":"journal-article","created":{"date-parts":[[2025,11,21]],"date-time":"2025-11-21T14:28:41Z","timestamp":1763735321000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["SWIR-LightFusion: multi-spectral semantic fusion of synthetic SWIR with thermal IR (LWIR\/MWIR) and RGB"],"prefix":"10.1007","volume":"29","author":[{"given":"Muhammad Ishfaq","family":"Hussain","sequence":"first","affiliation":[]},{"given":"Ma","family":"Van Linh","sequence":"additional","affiliation":[]},{"given":"Zubia","family":"Naz","sequence":"additional","affiliation":[]},{"given":"Unse","family":"Fatima","sequence":"additional","affiliation":[]},{"given":"Yeongmin","family":"Ko","sequence":"additional","affiliation":[]},{"given":"Moongu","family":"Jeon","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,21]]},"reference":[{"key":"5792_CR1","doi-asserted-by":"crossref","unstructured":"Liu, J., Fan, X., Huang, Z., Wu, G., Liu, R., Zhong, W., Luo, Z.: Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5802\u20135811 (2022)","DOI":"10.1109\/CVPR52688.2022.00571"},{"key":"5792_CR2","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","volume":"83","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Zhang, H., Jiang, X., Ma, J.: PIAFusion: A progressive infrared and visible image fusion network based on illumination aware. Information Fusion 83, 79\u201392 (2022)","journal-title":"Information Fusion"},{"key":"5792_CR3","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","volume":"76","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Xu, H., Tian, X., Jiang, J., Ma, J.: Image fusion meets deep learning: A survey and perspective. Information Fusion 76, 323\u2013336 (2021)","journal-title":"Information Fusion"},{"issue":"6","key":"5792_CR4","doi-asserted-by":"publisher","first-page":"5971","DOI":"10.1109\/TVT.2022.3161378","volume":"71","author":"MI Hussain","year":"2022","unstructured":"Hussain, M.I., Azam, S., Rafique, M.A., Sheri, A.M., Jeon, M.: Drivable region estimation for self-driving vehicles using radar. IEEE Trans. Veh. Technol. 71(6), 5971\u20135982 (2022)","journal-title":"IEEE Trans. Veh. Technol."},{"key":"5792_CR5","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Information Fusion 82, 28\u201342 (2022)","journal-title":"Information Fusion"},{"key":"5792_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101870","volume":"99","author":"L Tang","year":"2023","unstructured":"Tang, L., Zhang, H., Xu, H., Ma, J.: Rethinking the necessity of image fusion in high-level viseion tasks: A practical infrared and visible image fusion network based on progressive semantic injection and scene fidelity. Information Fusion 99, 101870 (2023)","journal-title":"Information Fusion"},{"key":"5792_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.128183","volume":"600","author":"J Liu","year":"2024","unstructured":"Liu, J., Zhang, Q.: Multi-level modality-specific and modality-common features fusion network for RGB-IR person re-identification. Neurocomputing 600, 128183 (2024)","journal-title":"Neurocomputing"},{"key":"5792_CR8","doi-asserted-by":"crossref","unstructured":"Hussain, M.I., Rafique, M.A., Khurbaev, S., Jeon, M.: Exploring data variance challenges in fusion of radar and camera for robotics and autonomous driving. In: 2022 10th International Conference on Control, Mechatronics and Automation (ICCMA), pp. 7\u201312 (2022). IEEE","DOI":"10.1109\/ICCMA56665.2022.10011604"},{"key":"5792_CR9","doi-asserted-by":"publisher","first-page":"1635","DOI":"10.1109\/TNSRE.2023.3254151","volume":"31","author":"MI Hussain","year":"2023","unstructured":"Hussain, M.I., Rafique, M.A., Kim, J., Jeon, M., Pedrycz, W.: Artificial proprioceptive reflex warning using EMG in advanced driving assistance system. IEEE Trans. Neural Syst. Rehabil. Eng. 31, 1635\u20131644 (2023)","journal-title":"IEEE Trans. Neural Syst. Rehabil. Eng."},{"key":"5792_CR10","unstructured":"Zhao, T., Yuan, M., Wei, X.: Removal and selection: Improving RGB-infrared object detection via coarse-to-fine fusion. arXiv preprint arXiv:2401.10731 (2024)"},{"key":"5792_CR11","doi-asserted-by":"crossref","unstructured":"Hussain, M.I., Azam, S., Munir, F., Khan, Z., Jeon, M.: Multiple objects tracking using radar for autonomous driving. In: 2020 IEEE International IOT, Electronics and Mechatronics Conference (IEMTRONICS), pp. 1\u20134 (2020)","DOI":"10.1109\/IEMTRONICS51293.2020.9216363"},{"key":"5792_CR12","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2024.110785","volume":"156","author":"LV Ma","year":"2024","unstructured":"Ma, L.V., Nguyen, T.T.D., Shim, C., Kim, D.Y., Ha, N., Jeon, M.: Visual multi-object tracking with re-identification and occlusion handling using labeled random finite sets. Pattern Recognition 156, 110785 (2024)","journal-title":"Pattern Recognition"},{"key":"5792_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102496","author":"LV Ma","year":"2024","unstructured":"Ma, L.V., Nguyen, T.T.D., Vo, B.-N., Jang, H., Jeon, M.: Track initialization and re-identification for 3D multi-view multi-object tracking. Information Fusion (2024). https:\/\/doi.org\/10.1016\/j.inffus.2024.102496","journal-title":"Information Fusion"},{"key":"5792_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.cmpb.2024.108122","volume":"248","author":"KR Ranipa","year":"2024","unstructured":"Ranipa, K.R., Zhu, W.-P., Swamy, M.N.S.: A novel feature-level fusion scheme with multimodal attention cnn for heart sound classification. Comput. Methods Programs Biomed 248, 108122 (2024)","journal-title":"Comput. Methods Programs Biomed"},{"key":"5792_CR15","doi-asserted-by":"crossref","unstructured":"Hussain, M.I., Naz, Z., Van\u00a0Ma, L., Gwak, J., Jeon, M.: A light gradient residual encoder-decoder network for multimodal image fusion. In: 2024 International Conference on Intelligent Computing, Communication, Networking and Services (ICCNS), pp. 1\u20136 (2024). IEEE","DOI":"10.1109\/ICCNS62192.2024.10776536"},{"key":"5792_CR16","doi-asserted-by":"crossref","unstructured":"Kumar, N., Upadhyay, A., Sharma, S., Sharma, M., Singh, P.: MWIRSTD: A mwir small target detection dataset. In: 2024 IEEE International Conference on Image Processing (ICIP), pp. 179\u2013185 (2024). IEEE","DOI":"10.1109\/ICIP51287.2024.10648074"},{"key":"5792_CR17","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1016\/j.dib.2017.09.038","volume":"15","author":"A Toet","year":"2017","unstructured":"Toet, A.: The TNO multiband image data collection. Data in Brief 15, 249\u2013251 (2017)","journal-title":"Data in Brief"},{"key":"5792_CR18","doi-asserted-by":"crossref","unstructured":"Gebhardt, E., Wolf, M.: CAMEL dataset for visual and thermal infrared multiple object detection and tracking. In: 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pp. 1\u20136 (2018). IEEE","DOI":"10.1109\/AVSS.2018.8639094"},{"key":"5792_CR19","doi-asserted-by":"crossref","unstructured":"Xu, H., Ma, J., Le, Z., Jiang, J., Guo, X.: Fusiondn: A unified densely connected network for image fusion. In: Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (2020)","DOI":"10.1609\/aaai.v34i07.6936"},{"issue":"7","key":"5792_CR20","doi-asserted-by":"publisher","first-page":"4090","DOI":"10.1007\/s00034-022-01975-y","volume":"41","author":"H-T Hu","year":"2022","unstructured":"Hu, H.-T., Lee, T.-T.: Robust complementary dual image watermarking in subbands derived from the laplacian pyramid, discrete wavelet transform, and directional filter bank. Circuits, Systems, and Signal Processing 41(7), 4090\u20134116 (2022)","journal-title":"Circuits, Systems, and Signal Processing"},{"issue":"2","key":"5792_CR21","doi-asserted-by":"publisher","first-page":"491","DOI":"10.1109\/TMI.2017.2760978","volume":"37","author":"J Schlemper","year":"2017","unstructured":"Schlemper, J., Caballero, J., Hajnal, J.V., Price, A.N., Rueckert, D.: A deep cascade of convolutional neural networks for dynamic MR image reconstruction. IEEE Trans. Med. Imaging 37(2), 491\u2013503 (2017)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"5792_CR22","doi-asserted-by":"publisher","first-page":"110","DOI":"10.1016\/j.inffus.2021.02.019","volume":"72","author":"Y Fu","year":"2021","unstructured":"Fu, Y., Wu, X., Durrani, T.S.: Image fusion based on generative adversarial network consistent with perception. Information Fusion 72, 110\u2013125 (2021)","journal-title":"Information Fusion"},{"key":"5792_CR23","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.127913","volume":"595","author":"P Li","year":"2024","unstructured":"Li, P., Chen, J., Lin, B., Xu, X.: Residual spatial fusion network for RGB-thermal semantic segmentation. Neurocomputing 595, 127913 (2024)","journal-title":"Neurocomputing"},{"key":"5792_CR24","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: FusionGAN: A generative adversarial network for infrared and visible image fusion. Information fusion 48, 11\u201326 (2019)","journal-title":"Information fusion"},{"key":"5792_CR25","doi-asserted-by":"crossref","unstructured":"Musa, P., Al\u00a0Rafi, F., Lamsani, M.: A review: Contrast-limited adaptive histogram equalization (CLAHE) methods to help the application of face recognition. In: 2018 Third International Conference on Informatics and Computing (ICIC), pp. 1\u20136 (2018). IEEE","DOI":"10.1109\/IAC.2018.8780492"},{"key":"5792_CR26","doi-asserted-by":"publisher","first-page":"12484","DOI":"10.1609\/aaai.v34i07.6936","volume":"34","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Le, Z., Jiang, J., Guo, X.: FusionDN: A unified densely connected network for image fusion. Proceedings of the AAAI Conference on Artificial Intelligence 34, 12484\u201312491 (2020)","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"issue":"3","key":"5792_CR27","doi-asserted-by":"publisher","first-page":"379","DOI":"10.1002\/j.1538-7305.1948.tb01338.x","volume":"27","author":"CE Shannon","year":"1948","unstructured":"Shannon, C.E.: A mathematical theory of communication. Bell Syst. Tech. J. 27(3), 379\u2013423 (1948)","journal-title":"Bell Syst. Tech. J."},{"issue":"7","key":"5792_CR28","doi-asserted-by":"publisher","first-page":"313","DOI":"10.1049\/el:20020212","volume":"38","author":"G Qu","year":"2002","unstructured":"Qu, G., Zhang, D., Yan, P.: Information measure for performance of image fusion. Electron. Lett. 38(7), 313\u2013315 (2002)","journal-title":"Electron. Lett."},{"key":"5792_CR29","unstructured":"Gonzalez, R.C., Woods, R.E.: Digital Image Processing. Prentice Hall, ??? (2002)"},{"issue":"12","key":"5792_CR30","doi-asserted-by":"publisher","first-page":"2959","DOI":"10.1109\/26.477498","volume":"43","author":"AM Eskicioglu","year":"1995","unstructured":"Eskicioglu, A.M., Fisher, P.S.: Image quality measures and their performance. IEEE Trans. Commun. 43(12), 2959\u20132965 (1995)","journal-title":"IEEE Trans. Commun."},{"issue":"4","key":"5792_CR31","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"issue":"4","key":"5792_CR32","doi-asserted-by":"publisher","first-page":"308","DOI":"10.1049\/el:20000267","volume":"36","author":"CS Xydeas","year":"2000","unstructured":"Xydeas, C.S., Petrovic, V.: Objective image fusion performance measure. Electron. Lett. 36(4), 308\u2013309 (2000)","journal-title":"Electron. Lett."},{"issue":"2","key":"5792_CR33","doi-asserted-by":"publisher","first-page":"430","DOI":"10.1109\/TIP.2005.859378","volume":"15","author":"HR Sheikh","year":"2006","unstructured":"Sheikh, H.R., Bovik, A.C.: Image information and visual quality. IEEE Transactions on Image Processing 15(2), 430\u2013444 (2006)","journal-title":"IEEE Transactions on Image Processing"},{"key":"5792_CR34","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Bai, H., Zhang, J., Zhang, Y., Xu, S., Lin, Z., Timofte, R., Gool, L.V.: CDDFuse: Correlation-driven dual-branch feature decomposition for multi-modality image fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5906\u20135916 (2023)","DOI":"10.1109\/CVPR52729.2023.00572"},{"key":"5792_CR35","doi-asserted-by":"publisher","first-page":"1028","DOI":"10.1609\/aaai.v33i01.33011028","volume":"33","author":"A Liu","year":"2019","unstructured":"Liu, A., Liu, X., Fan, J., Ma, Y., Zhang, A., Xie, H., Tao, D.: Perceptual-sensitive gan for generating adversarial patches. Proceedings of the AAAI Conference on Artificial Intelligence 33, 1028\u20131035 (2019)","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"key":"5792_CR36","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Xu, S., Zhang, C., Liu, J., Li, P., Zhang, J.: DIDFuse: Deep image decomposition for infrared and visible image fusion. In: International Joint Conference on Artificial Intelligence (2020)","DOI":"10.24963\/ijcai.2020\/135"},{"issue":"10","key":"5792_CR37","doi-asserted-by":"publisher","first-page":"2761","DOI":"10.1007\/s11263-021-01501-8","volume":"129","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Ma, J.: SDNet: A versatile squeeze-and-decomposition network for real-time image fusion. International Journal of Computer Vision 129(10), 2761\u20132785 (2021)","journal-title":"International Journal of Computer Vision"},{"key":"5792_CR38","doi-asserted-by":"crossref","unstructured":"Chen, H., Wang, Y., Guo, T., Xu, C., Deng, Y., Liu, Z., Ma, S., Xu, C., Xu, C., Gao, W.: Pre-trained image processing transformer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12299\u201312310 (2021)","DOI":"10.1109\/CVPR46437.2021.01212"},{"key":"5792_CR39","doi-asserted-by":"crossref","unstructured":"Xu, H., Ma, J., Yuan, J., Le, Z., Liu, W.: RFNet: Unsupervised network for mutually reinforcing multi-modal image registration and fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 19679\u201319688 (2022)","DOI":"10.1109\/CVPR52688.2022.01906"},{"issue":"1","key":"5792_CR40","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TPAMI.2020.3012548","volume":"44","author":"H Xu","year":"2020","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 502\u2013518 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"5792_CR41","first-page":"1","volume":"70","author":"J Ma","year":"2020","unstructured":"Ma, J., Zhang, H., Shao, Z., Liang, P., Xu, H.: Ganmcc: A generative adversarial network with multiclassification constraints for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 70, 1\u201314 (2020)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"5792_CR42","doi-asserted-by":"crossref","unstructured":"Wang, D., Liu, J., Fan, X., Liu, R.: Unsupervised misaligned infrared and visible image fusion via cross-modality image generation and registration. arXiv preprint arXiv:2205.11876 (2022)","DOI":"10.24963\/ijcai.2022\/487"},{"key":"5792_CR43","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Bai, H., Zhu, Y., Zhang, J., Xu, S., Zhang, Y., Zhang, K., Meng, D., Timofte, R., Van\u00a0Gool, L.: Ddfm: denoising diffusion model for multi-modality image fusion. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8082\u20138093 (2023)","DOI":"10.1109\/ICCV51070.2023.00742"},{"issue":"4","key":"5792_CR44","doi-asserted-by":"publisher","first-page":"532","DOI":"10.1109\/TCOM.1983.1095851","volume":"31","author":"PJ Burt","year":"1983","unstructured":"Burt, P.J., Adelson, E.H.: The laplacian pyramid as a compact image code. IEEE Trans. Commun. 31(4), 532\u2013540 (1983)","journal-title":"IEEE Trans. Commun."},{"key":"5792_CR45","unstructured":"Li, H., Wu, X.-J., Kittler, J.: Mdlatlrr: A novel decomposition method for infrared and visible image fusion. arXiv:1811.02291 (2018)"},{"issue":"7","key":"5792_CR46","doi-asserted-by":"publisher","first-page":"2864","DOI":"10.1109\/TIP.2013.2244222","volume":"22","author":"S Li","year":"2013","unstructured":"Li, S., Kang, X., Hu, J.: Image fusion with guided filtering. IEEE Trans. Image Process. 22(7), 2864\u20132875 (2013)","journal-title":"IEEE Trans. Image Process."},{"issue":"7","key":"5792_CR47","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma, J., Tang, L., Fan, F., Huang, J., Mei, X., Ma, Y.: Swinfusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA Journal of Automatica Sinica 9(7), 1200\u20131217 (2022)","journal-title":"IEEE\/CAA Journal of Automatica Sinica"},{"key":"5792_CR48","doi-asserted-by":"crossref","unstructured":"Edstedt, J., Sun, Q., B\u00f6kman, G., Wadenb\u00e4ck, M., Felsberg, M.: RoMa: Robust Dense Feature Matching. IEEE Conference on Computer Vision and Pattern Recognition (2024)","DOI":"10.1109\/CVPR52733.2024.01871"}],"container-title":["Cluster Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-025-05792-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10586-025-05792-1","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-025-05792-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T13:07:43Z","timestamp":1773925663000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10586-025-05792-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,21]]},"references-count":48,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["5792"],"URL":"https:\/\/doi.org\/10.1007\/s10586-025-05792-1","relation":{},"ISSN":["1386-7857","1573-7543"],"issn-type":[{"value":"1386-7857","type":"print"},{"value":"1573-7543","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,21]]},"assertion":[{"value":"31 March 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 September 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 October 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 November 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}},{"value":"The authors declare no competing interests.\u00a0Not applicable. This study did not involve human participants, human data, or animals. It utilized one private dataset collected without human or animal involvement, as well as publicly available open-source datasets; therefore, no ethics approval was required.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval statement"}}],"article-number":"48"}}