{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T16:34:50Z","timestamp":1773246890979,"version":"3.50.1"},"reference-count":59,"publisher":"MDPI AG","issue":"1","license":[{"start":{"date-parts":[[2023,12,30]],"date-time":"2023-12-30T00:00:00Z","timestamp":1703894400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"project CEIEC-2022-ZM02-0247"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Multiple object tracking (MOT) plays an important role in intelligent video-processing tasks, which aims to detect and track all moving objects in a scene. Joint-detection-and-tracking (JDT) methods are thriving in MOT tasks, because they accomplish the detection and data association in a single stage. However, the slow training convergence and insufficient data association limit the performance of JDT methods. In this paper, the anchor-based query (ABQ) is proposed to improve the design of the JDT methods for faster training convergence. By augmenting the coordinates of the anchor boxes into the learnable queries of the decoder, the ABQ introduces explicit prior spatial knowledge into the queries to focus the query-to-feature learning of the JDT methods on the local region, which leads to faster training speed and better performance. Moreover, a new template matching (TM) module is designed for the JDT methods, which enables the JDT methods to associate the detection results and trajectories with historical features. Finally, a new transformer-based MOT method, ABQ-Track, is proposed. Extensive experiments verify the effectiveness of the two modules, and the ABQ-Track surpasses the performance of the baseline JDT methods, TransTrack. Specifically, the ABQ-Track only needs to train for 50 epochs to achieve convergence, while that for TransTrack is 150 epochs.<\/jats:p>","DOI":"10.3390\/s24010229","type":"journal-article","created":{"date-parts":[[2023,12,31]],"date-time":"2023-12-31T06:00:21Z","timestamp":1704002421000},"page":"229","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["Transformer-Based Multiple-Object Tracking via Anchor-Based-Query and Template Matching"],"prefix":"10.3390","volume":"24","author":[{"given":"Qinyu","family":"Wang","sequence":"first","affiliation":[{"name":"State Key Laboratory of Integrated Service Networks, School of Telecommunications Engineering, Xidian University, No. 2, South Taibai Street, Hi-Tech Development Zone, Xi\u2019an 710071, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-4275-3037","authenticated-orcid":false,"given":"Chenxu","family":"Lu","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Integrated Service Networks, School of Telecommunications Engineering, Xidian University, No. 2, South Taibai Street, Hi-Tech Development Zone, Xi\u2019an 710071, China"}]},{"given":"Long","family":"Gao","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Integrated Service Networks, School of Telecommunications Engineering, Xidian University, No. 2, South Taibai Street, Hi-Tech Development Zone, Xi\u2019an 710071, China"}]},{"given":"Gang","family":"He","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Integrated Service Networks, School of Telecommunications Engineering, Xidian University, No. 2, South Taibai Street, Hi-Tech Development Zone, Xi\u2019an 710071, China"}]}],"member":"1968","published-online":{"date-parts":[[2023,12,30]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Cheng, C.C., Qiu, M.X., Chiang, C.K., and Lai, S.H. (2023, January 2\u20136). ReST: A Reconfigurable Spatial-Temporal Graph Model for Multi-Camera Multi-Object Tracking. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Paris, France.","DOI":"10.1109\/ICCV51070.2023.00922"},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Pang, J., Qiu, L., Li, X., Chen, H., Li, Q., Darrell, T., and Yu, F. (2021, January 19\u201325). Quasi-dense similarity learning for multiple object tracking. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Online.","DOI":"10.1109\/CVPR46437.2021.00023"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Sun, P., Jiang, Y., Yu, D., Weng, F., Yuan, Z., Luo, P., Liu, W., and Wang, X. (2022, January 23\u201327). Bytetrack: Multi-object tracking by associating every detection box. Proceedings of the European Conference on Computer Vision, Tel-Aviv, Israel.","DOI":"10.1007\/978-3-031-20047-2_1"},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, K., Oh, S.W., Kweon, I.S., and Lee, J.Y. (2022, January 23\u201327). Tracking by Associating Clips. Proceedings of the European Conference on Computer Vision, Tel-Aviv, Israel.","DOI":"10.1007\/978-3-031-19806-9_8"},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Zhou, X., Yin, T., Koltun, V., and Kr\u00e4henb\u00fchl, P. (2022, January 19\u201324). Global tracking transformers. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.00857"},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Kieritz, H., Hubner, W., and Arens, M. (2018, January 18\u201322). Joint detection and online multi-object tracking. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPRW.2018.00195"},{"key":"ref_7","unstructured":"Sun, P., Cao, J., Jiang, Y., Zhang, R., Xie, E., Yuan, Z., Wang, C., and Luo, P. (2020). Transtrack: Multiple object tracking with transformer. arXiv."},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Meinhardt, T., Kirillov, A., Leal-Taixe, L., and Feichtenhofer, C. (2022, January 19\u201324). Trackformer: Multi-object tracking with transformers. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.00864"},{"key":"ref_9","unstructured":"Liu, S., Li, F., Zhang, H., Yang, X., Qi, X., Su, H., Zhu, J., and Zhang, L. (2022, January 25\u201329). DAB-DETR: Dynamic Anchor Boxes are Better Queries for DETR. Proceedings of the International Conference on Learning Representations, Online."},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Meng, D., Chen, X., Fan, Z., Zeng, G., Li, H., Yuan, Y., Sun, L., and Wang, J. (2021, January 10\u201317). Conditional detr for fast training convergence. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, QC, Canada.","DOI":"10.1109\/ICCV48922.2021.00363"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Cai, J., Xu, M., Li, W., Xiong, Y., Xia, W., Tu, Z., and Soatto, S. (2022, January 19\u201324). Memot: Multi-object tracking with memory. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.00792"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Li, S., Kong, Y., and Rezatofighi, H. (2022, January 19\u201324). Learning of global objective for network flow in multi-object tracking. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.00865"},{"key":"ref_13","unstructured":"Milan, A., Leal-Taix\u00e9, L., Reid, I., Roth, S., and Schindler, K. (2016). MOT16: A benchmark for multi-object tracking. arXiv."},{"key":"ref_14","unstructured":"Dendorfer, P., Rezatofighi, H., Milan, A., Shi, J., Cremers, D., Reid, I., Roth, S., Schindler, K., and Leal-Taix\u00e9, L. (2020). Mot20: A benchmark for multi object tracking in crowded scenes. arXiv."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Dave, A., Khurana, T., Tokmakov, P., Schmid, C., and Ramanan, D. (2020, January 23\u201328). Tao: A large-scale benchmark for tracking any object. Proceedings of the European Conference on Computer Vision, Online.","DOI":"10.1007\/978-3-030-58558-7_26"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., and Belongie, S. (2017, January 21\u201326). Feature pyramid networks for object detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Hawaii, HI, USA.","DOI":"10.1109\/CVPR.2017.106"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., and Doll\u00e1r, P. (2017, January 22\u201329). Focal loss for dense object detection. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.324"},{"key":"ref_18","unstructured":"Ren, S., He, K., Girshick, R., and Sun, J. (2021, January 8\u201312). Faster r-cnn: Towards real-time object detection with region proposal networks. Proceedings of the 28th International Conference on Neural Information Processing Systems, Plaza Hotel Bali, Indonesia."},{"key":"ref_19","unstructured":"Redmon, J., and Farhadi, A. (2018). Yolov3: An incremental improvement. arXiv."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Bewley, A., Ge, Z., Ott, L., Ramos, F., and Upcroft, B. (2016, January 25\u201328). Simple online and realtime tracking. Proceedings of the 2016 IEEE International Conference on Image Processing, Phoenix, AZ, USA.","DOI":"10.1109\/ICIP.2016.7533003"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Wojke, N., Bewley, A., and Paulus, D. (2017, January 17\u201320). Simple online and realtime tracking with a deep association metric. Proceedings of the 2017 IEEE International Conference on Image Processing, Beijing, China.","DOI":"10.1109\/ICIP.2017.8296962"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Wang, Z., Zheng, L., Liu, Y., Li, Y., and Wang, S. (2020, January 23\u201328). Towards real-time multi-object tracking. Proceedings of the European Conference on Computer Vision, Online.","DOI":"10.1007\/978-3-030-58621-8_7"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"8725","DOI":"10.1109\/TMM.2023.3240881","article-title":"Strongsort: Make deepsort great again","volume":"25","author":"Du","year":"2023","journal-title":"IEEE Trans. Multimed."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Du, Y., Wan, J., Zhao, Y., Zhang, B., Tong, Z., and Dong, J. (2021, January 10\u201317). Giaotracker: A comprehensive framework for mcmot with global information and optimizing strategies in visdrone 2021. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, QC, Canada.","DOI":"10.1109\/ICCVW54120.2021.00315"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"75","DOI":"10.1016\/j.neucom.2021.12.104","article-title":"Mat: Motion-aware multi-object tracking","volume":"476","author":"Han","year":"2022","journal-title":"Neurocomputing"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Stadler, D., and Beyerer, J. (2022, January 4\u20138). Modelling ambiguous assignments for multi-person tracking in crowds. Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA.","DOI":"10.1109\/WACVW54805.2022.00019"},{"key":"ref_27","unstructured":"Aharon, N., Orfaig, R., and Bobrovsky, B.Z. (2022). BoT-SORT: Robust associations multi-pedestrian tracking. arXiv."},{"key":"ref_28","unstructured":"Leal-Taix\u00e9, L., Canton-Ferrer, C., and Schindler, K. (July, January 26). Learning by tracking: Siamese CNN for robust target association. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Las Vegas, NV, USA."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Ristani, E., and Tomasi, C. (2018, January 30\u201331). Features for multi-target multi-camera tracking and re-identification. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Istanbul, Turkey.","DOI":"10.1109\/CVPR.2018.00632"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Pinz, A., and Zisserman, A. (2017, January 22\u201329). Detect to track and track to detect. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.330"},{"key":"ref_31","unstructured":"Bergmann, P., Meinhardt, T., and Leal-Taixe, L. (November, January 27). Tracking without bells and whistles. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Seoul, Republic of Korea."},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Zhou, X., Koltun, V., and Kr\u00e4henb\u00fchl, P. (2020, January 23\u201328). Tracking objects as points. Proceedings of the European Conference on Computer Vision, Online.","DOI":"10.1007\/978-3-030-58548-8_28"},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"7892","DOI":"10.1109\/JIOT.2020.2996609","article-title":"Multiplex labeling graph for near-online tracking in crowded scenes","volume":"7","author":"Zhang","year":"2020","journal-title":"IEEE Internet Things J."},{"key":"ref_34","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., and Polosukhin, I. (2017, January 4\u20139). Attention is all you need. Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA."},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Wu, H., Xiao, B., Codella, N., Liu, M., Dai, X., Yuan, L., and Zhang, L. (2021, January 10\u201317). Cvt: Introducing convolutions to vision transformers. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, QC, Canada.","DOI":"10.1109\/ICCV48922.2021.00009"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., and Guo, B. (2021, January 10\u201317). Swin transformer: Hierarchical vision transformer using shifted windows. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, QC, Canada.","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., and Zagoruyko, S. (2020, January 23\u201328). End-to-end object detection with transformers. Proceedings of the European Conference on Computer Vision, Online.","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Li, Y., Mao, H., Girshick, R., and He, K. (2022, January 23\u201327). Exploring plain vision transformer backbones for object detection. Proceedings of the European Conference on Computer Vision, Tel-Aviv, Israel.","DOI":"10.1007\/978-3-031-20077-9_17"},{"key":"ref_39","doi-asserted-by":"crossref","unstructured":"Wang, Y., Xu, Z., Wang, X., Shen, C., Cheng, B., Shen, H., and Xia, H. (2021, January 19\u201325). End-to-end video instance segmentation with transformers. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Online.","DOI":"10.1109\/CVPR46437.2021.00863"},{"key":"ref_40","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., and Gelly, S. (2020). An image is worth 16x16 words: Transformers for image recognition at scale. arXiv."},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Wang, W., Xie, E., Li, X., Fan, D.P., Song, K., Liang, D., Lu, T., Luo, P., and Shao, L. (2021, January 10\u201317). Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. Proceedings of the IEEE\/CVF international conference on computer vision, Montreal, QC, Canada.","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"Chen, Z., Zhu, Y., Zhao, C., Hu, G., Zeng, W., Wang, J., and Tang, M. (2021, January 20\u201324). Dpt: Deformable patch-based transformer for visual recognition. Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China.","DOI":"10.1145\/3474085.3475467"},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Dong, X., Bao, J., Chen, D., Zhang, W., Yu, N., Yuan, L., Chen, D., and Guo, B. (2022, January 19\u201324). Cswin transformer: A general vision transformer backbone with cross-shaped windows. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.01181"},{"key":"ref_44","doi-asserted-by":"crossref","unstructured":"Tu, Z., Talebi, H., Zhang, H., Yang, F., Milanfar, P., Bovik, A., and Li, Y. (2022, January 23\u201327). Maxvit: Multi-axis vision transformer. Proceedings of the European Conference on Computer Vision, Tel-Aviv, Israel.","DOI":"10.1007\/978-3-031-20053-3_27"},{"key":"ref_45","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., and Dai, J. (2021, January 3\u20137). Deformable detr: Deformable transformers for end-to-end object detection. Proceedings of the 9th International Conference on Learning Representations, Virtual Event, Austria."},{"key":"ref_46","doi-asserted-by":"crossref","unstructured":"Dai, X., Chen, Y., Yang, J., Zhang, P., Yuan, L., and Zhang, L. (2021, January 10\u201317). Dynamic detr: End-to-end object detection with dynamic attention. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, QC, Canada.","DOI":"10.1109\/ICCV48922.2021.00298"},{"key":"ref_47","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (July, January 26). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA."},{"key":"ref_48","unstructured":"Shao, S., Zhao, Z., Li, B., Xiao, T., Yu, G., Zhang, X., and Sun, J. (2018). Crowdhuman: A benchmark for detecting human in a crowd. arXiv."},{"key":"ref_49","doi-asserted-by":"crossref","unstructured":"Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., and Savarese, S. (2019, January 16\u201320). Generalized intersection over union: A metric and a loss for bounding box regression. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00075"},{"key":"ref_50","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., and Fei-Fei, L. (2009, January 20\u201325). Imagenet: A large-scale hierarchical image database. Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref_51","unstructured":"Glorot, X., and Bengio, Y. (2010, January 13\u201315). Understanding the difficulty of training deep feedforward neural networks. Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Sardinia, Italy."},{"key":"ref_52","doi-asserted-by":"crossref","unstructured":"Yin, J., Wang, W., Meng, Q., Yang, R., and Shen, J. (2020, January 14\u201319). A unified object motion and affinity model for online multi-object tracking. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00680"},{"key":"ref_53","doi-asserted-by":"crossref","unstructured":"Pang, B., Li, Y., Zhang, Y., Li, M., and Lu, C. (2020, January 14\u201319). Tubetk: Adopting tubes to track multi-object in a one-step training model. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00634"},{"key":"ref_54","doi-asserted-by":"crossref","unstructured":"Wu, J., Cao, J., Song, L., Wang, Y., Yang, M., and Yuan, J. (2021, January 19\u201325). Track to detect and segment: An online multi-object tracker. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Online.","DOI":"10.1109\/CVPR46437.2021.01217"},{"key":"ref_55","doi-asserted-by":"crossref","unstructured":"Zheng, L., Tang, M., Chen, Y., Zhu, G., Wang, J., and Lu, H. (2021, January 19\u201325). Improving multiple object tracking with single object tracking. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Online.","DOI":"10.1109\/CVPR46437.2021.00248"},{"key":"ref_56","first-page":"7820","article-title":"Transcenter: Transformers with dense queries for multiple-object tracking","volume":"15","author":"Xu","year":"2021","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_57","first-page":"1","article-title":"A simple baseline for multi-object tracking","volume":"7","author":"Zhang","year":"2021","journal-title":"Int. J. Comput. Vis."},{"key":"ref_58","doi-asserted-by":"crossref","first-page":"3182","DOI":"10.1109\/TIP.2022.3165376","article-title":"Rethinking the competition between detection and reid in multiobject tracking","volume":"31","author":"Liang","year":"2022","journal-title":"IEEE Trans. Image Process."},{"key":"ref_59","doi-asserted-by":"crossref","unstructured":"Wang, Q., Zheng, Y., Pan, P., and Xu, Y. (2021, January 19\u201325). Multiple object tracking with correlation learning. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Online.","DOI":"10.1109\/CVPR46437.2021.00387"}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/24\/1\/229\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T21:45:03Z","timestamp":1760132703000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/24\/1\/229"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,30]]},"references-count":59,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2024,1]]}},"alternative-id":["s24010229"],"URL":"https:\/\/doi.org\/10.3390\/s24010229","relation":{},"ISSN":["1424-8220"],"issn-type":[{"value":"1424-8220","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12,30]]}}}