{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T10:30:39Z","timestamp":1761388239322,"version":"build-2065373602"},"reference-count":56,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T00:00:00Z","timestamp":1755734400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T00:00:00Z","timestamp":1755734400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the Gansu Provincial Key Laboratory of Advanced Control of Industrial Process Open Fund Project","award":["2022KX10"],"award-info":[{"award-number":["2022KX10"]}]},{"name":"the Natural Science Foundation of Gansu Province","award":["23JRRA796"],"award-info":[{"award-number":["23JRRA796"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62061042"],"award-info":[{"award-number":["62061042"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s00530-025-01913-3","type":"journal-article","created":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T11:41:08Z","timestamp":1755776468000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["VAFTrack: asynchronous feature fusion via visual receptive weighted key-value perceptual for visual tracking"],"prefix":"10.1007","volume":"31","author":[{"given":"Zhongmin","family":"Liu","sequence":"first","affiliation":[]},{"given":"Zhenhua","family":"Li","sequence":"additional","affiliation":[]},{"given":"Wenjin","family":"Hu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,21]]},"reference":[{"issue":"7","key":"1913_CR1","first-page":"109","volume":"44","author":"J Chen","year":"2024","unstructured":"Chen, J., Li, D., Zeng, X., et al.: Cross-Modal optical information interaction and template dynamic update for RGBT target tracking method [J]. Acta Optica Sinica. 44(7), 109\u2013123 (2024)","journal-title":"Acta Optica Sinica"},{"key":"1913_CR2","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition[C]\/\/Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"1913_CR3","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I.: Attention is all you need. Adv. Neural. Inf. Process. Syst., 5998\u20136008 (2017)"},{"key":"1913_CR4","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., et al.: Going deeper with convolutions[C]\/\/Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 1\u20139 (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"1913_CR5","doi-asserted-by":"crossref","unstructured":"Chen, X., Yan, B., Zhu, J., et al.: Transformer tracking[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 8126\u20138135 (2021)","DOI":"10.1109\/CVPR46437.2021.00803"},{"key":"1913_CR6","doi-asserted-by":"crossref","unstructured":"Yan, B., Peng, H., Fu, J., et al.: Learning spatio-temporal transformer for visual tracking[C]\/\/Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 10448\u201310457 (2021)","DOI":"10.1109\/ICCV48922.2021.01028"},{"key":"1913_CR7","doi-asserted-by":"crossref","unstructured":"Wang, N., Zhou, W., Wang, J., et al.: Transformer meets tracker: Exploiting temporal context for robust visual tracking[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 1571\u20131580 (2021)","DOI":"10.1109\/CVPR46437.2021.00162"},{"key":"1913_CR8","doi-asserted-by":"crossref","unstructured":"Cui, Y., Jiang, C., Wang, L., et al.: Mixformer: end-to-end tracking with iterative mixed attention[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 13608\u201313618 (2022)","DOI":"10.1109\/CVPR52688.2022.01324"},{"key":"1913_CR9","doi-asserted-by":"crossref","unstructured":"Ye, B., Chang, H., Ma, B., et al.: Joint feature learning and relation modeling for tracking: a one-stream framework[C]\/\/European Conference on Computer Vision. Cham: Springer Nature, Switzerland, pp. 341\u2013357 (2022)","DOI":"10.1007\/978-3-031-20047-2_20"},{"issue":"2","key":"1913_CR10","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1049\/cvi2.12148","volume":"17","author":"S Qiu","year":"2023","unstructured":"Qiu, S., Gu, Y., Chen, M., et al.: A dynamic Adjust-Head Siamese network for object tracking. IET Comput. Vision. 17(2), 203\u2013210 (2023)","journal-title":"IET Comput. Vision"},{"key":"1913_CR11","doi-asserted-by":"crossref","unstructured":"Bhat, G., Danelljan, M., Gool, L.V., et al.: Learning discriminative model prediction for tracking[C]\/\/Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 6182\u20136191 (2019)","DOI":"10.1109\/ICCV.2019.00628"},{"key":"1913_CR12","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Bhat, G., Shahbaz Khan, F., et al.: ECO: efficient convolution operators for tracking[C]\/\/Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 6638\u20136646 (2017)","DOI":"10.1109\/CVPR.2017.733"},{"key":"1913_CR13","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Bhat, G., Khan, F.S., et al.: ATOM: accurate tracking by overlap maximization[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 4660\u20134669 (2019)","DOI":"10.1109\/CVPR.2019.00479"},{"key":"1913_CR14","doi-asserted-by":"publisher","first-page":"15095","DOI":"10.1007\/s11042-018-6562-8","volume":"79","author":"J Zhang","year":"2020","unstructured":"Zhang, J., Jin, X., Sun, J.: Spatial and semantic convolutional features for robust visual object tracking. Multimedia Tools Appl. 79, 15095\u201315115 (2020)","journal-title":"Multimedia Tools Appl"},{"key":"1913_CR15","first-page":"15180","volume":"2021","author":"B Yan","year":"2020","unstructured":"Yan, B., Peng, H., Wu, K., et al.: Lighttrack: finding lightweight neural networks for object tracking via one-shot architecture search. Proceed. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit. 2021, 15180\u201315189 (2020)","journal-title":"Proceed. IEEE\/CVF Conf. Comput. Vis. Pattern Recognit."},{"key":"1913_CR16","doi-asserted-by":"crossref","unstructured":"Borsuk, V., Vei, R., Kupyn, O., et al.: FEAR: fast, efficient, accurate and robust visual tracker[C]\/\/European Conference on Computer Vision, pp. 644\u2013663. Springer Nature Switzerland, Cham (2022)","DOI":"10.1007\/978-3-031-20047-2_37"},{"key":"1913_CR17","doi-asserted-by":"publisher","first-page":"107730","DOI":"10.1016\/j.compeleceng.2022.107730","volume":"98","author":"J Zhang","year":"2022","unstructured":"Zhang, J., Sun, J., Wang, J., et al.: An object tracking framework with recapture based on correlation filters and Siamese networks. Comput. Electr. Eng. 98, 107730 (2022)","journal-title":"Comput. Electr. Eng."},{"key":"1913_CR18","doi-asserted-by":"crossref","unstructured":"Kang, B., Chen, X., Wang, D., et al.: Exploring lightweight hierarchical vision transformers for efficient visual tracking[C]\/\/Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 9612\u20139621 (2023)","DOI":"10.1109\/ICCV51070.2023.00881"},{"key":"1913_CR19","unstructured":"Cui, Y., Song, T., Wu, G., et al.: Mixformerv2: efficient fully transformer tracking. Adv. Neural. Inf. Process. Syst., 36 (2024)"},{"key":"1913_CR20","doi-asserted-by":"crossref","unstructured":"Yang, X., Zeng, D., Wang, X., et al.: Adaptively bypassing vision transformer blocks for efficient visual Tracking. arxiv preprint arxiv:2406.08037 (2024)","DOI":"10.1016\/j.patcog.2024.111278"},{"key":"1913_CR21","doi-asserted-by":"crossref","unstructured":"Wei, Q., Zeng, B., Liu, J., et al.: Litetrack: layer pruning with asynchronous feature extraction for lightweight and efficient visual tracking[C]\/\/2024 IEEE International Conference on Robotics and Automation (ICRA). pp. 4968\u20134975. IEEE  (2024)","DOI":"10.1109\/ICRA57147.2024.10610022"},{"key":"1913_CR22","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2023.3309358","volume":"72","author":"X Chai","year":"2023","unstructured":"Chai, X., Shao, F., Chen, H., et al.: Super-resolution reconstruction for stereoscopic omnidirectional display systems via dynamic convolutions and cross-view transformer. IEEE Trans. Instrum. Meas. 72, 1\u201312 (2023)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"1913_CR23","doi-asserted-by":"publisher","first-page":"110625","DOI":"10.1016\/j.knosys.2023.110625","volume":"274","author":"X Chai","year":"2023","unstructured":"Chai, X., Shao, F., Jiang, Q., et al.: TCCL-Net: transformer-convolution collaborative learning network for omnidirectional image super-resolution. Knowl. Based Syst. 274, 110625 (2023)","journal-title":"Knowl. Based Syst."},{"key":"1913_CR24","doi-asserted-by":"crossref","unstructured":"Gao, J., Chen, M., Xu, C.: Learning probabilistic presence-absence evidence for weakly-supervised audio-visual event perception. IEEE Trans. Pattern Anal. Mach. Intell. (2025)","DOI":"10.1109\/TPAMI.2025.3546312"},{"key":"1913_CR25","unstructured":"Gao, J., Chen, M., Xiang, L., et al.: A comprehensive survey on evidential deep learning and its applications. arXiv preprint arXiv:2409.04720 (2024)"},{"key":"1913_CR26","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"1913_CR27","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., et al.: End-to-end object detection with transformers[C]\/\/European Conference on Computer Vision, pp. 213\u2013229. Springer International Publishing, Cham (2020)","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"1913_CR28","doi-asserted-by":"crossref","unstructured":"Mayer, C., Danelljan, M., Bhat, G., et al.: Transforming model prediction for tracking[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 8731\u20138740 (2022)","DOI":"10.1109\/CVPR52688.2022.00853"},{"issue":"9","key":"1913_CR29","doi-asserted-by":"publisher","first-page":"6186","DOI":"10.1109\/TCSVT.2022.3162599","volume":"32","author":"J Nie","year":"2022","unstructured":"Nie, J., Wu, H., He, Z., et al.: Spreading fine-grained prior knowledge for accurate tracking. IEEE Trans. Circuits Syst. Video Technol. 32(9), 6186\u20136199 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1913_CR30","unstructured":"Zhu, J., Chen, X., Diao, H., et al.: Exploring dynamic transformer for efficient object Tracking. arXiv preprint arXiv:2403.17651 (2024)"},{"key":"1913_CR31","doi-asserted-by":"publisher","first-page":"109075","DOI":"10.1016\/j.compeleceng.2024.109075","volume":"114","author":"J Zhang","year":"2024","unstructured":"Zhang, J., He, Y., Chen, W., et al.: CorrFormer: Context-aware tracking with cross-correlation and transformer. Comput. Electr. Eng. 114, 109075 (2024)","journal-title":"Comput. Electr. Eng."},{"key":"1913_CR32","doi-asserted-by":"crossref","unstructured":"Zhang, J., Qin, Y., Fan, S., et al.: SiamTFA: siamese triple-stream feature aggregation network for efficient RGBT tracking. IEEE Trans. Intell. Transp. Syst. (2024)","DOI":"10.1109\/TITS.2024.3512551"},{"issue":"1","key":"1913_CR33","doi-asserted-by":"publisher","first-page":"12256","DOI":"10.1038\/s41598-024-63028-5","volume":"14","author":"F Pan","year":"2024","unstructured":"Pan, F., Zhao, L., Wang, C.: Adaptive sparse attention-based compact transformer for object tracking. Sci. Rep. 14(1), 12256 (2024)","journal-title":"Sci. Rep."},{"key":"1913_CR34","doi-asserted-by":"publisher","first-page":"128908","DOI":"10.1016\/j.neucom.2024.128908","volume":"616","author":"J Zhang","year":"2025","unstructured":"Zhang, J., Yang, J., Liu, Z., et al.: Rgbt tracking via frequency-aware feature enhancement and unidirectional mixed attention. Neurocomputing. 616, 128908 (2025)","journal-title":"Neurocomputing"},{"key":"1913_CR35","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., et al.: Swin transformer: hierarchical vision transformer using shifted windows[C]\/\/Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"1913_CR36","first-page":"16743","volume":"35","author":"L Lin","year":"2022","unstructured":"Lin, L., Fan, H., Zhang, Z., et al.: SwinTrack: a simple and strong baseline for transformer tracking. Adv. Neural. Inf. Process. Syst. 35, 16743\u201316754 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1913_CR37","doi-asserted-by":"crossref","unstructured":"Song, Z., Yu, J., Chen, Y.P.P., et al.: Transformer tracking with cyclic shifting window attention[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 8791\u20138800 (2022)","DOI":"10.1109\/CVPR52688.2022.00859"},{"key":"1913_CR38","unstructured":"Xie, F., Wang, C., Wang, G.: Correlation-aware deep tracking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8751\u20138760"},{"key":"1913_CR39","first-page":"22","volume":"2021","author":"H Wu","year":"2022","unstructured":"Wu, H., Xiao, B., Codella, N., et al.: CvT: introducing convolutions to vision transformers. Proceed. IEEE\/CVF Int. Conf. Comput. Vis. 2021, 22\u201331 (2022)","journal-title":"Proceed. IEEE\/CVF Int. Conf. Comput. Vis."},{"issue":"9","key":"1913_CR40","doi-asserted-by":"publisher","first-page":"5102","DOI":"10.1109\/TCSVT.2023.3249468","volume":"33","author":"C Tang","year":"2023","unstructured":"Tang, C., Wang, X., Bai, Y., et al.: Learning spatial-frequency transformer for visual object tracking. IEEE Trans. Circuits Syst. Video Technol. 33(9), 5102\u20135116 (2023)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1913_CR41","doi-asserted-by":"publisher","first-page":"108485","DOI":"10.1016\/j.asoc.2022.108485","volume":"118","author":"J Zhang","year":"2022","unstructured":"Zhang, J., Feng, W., Yuan, T., et al.: SCSTCF: spatial-channel selection and Temporal regularized correlation filters for visual tracking. Appl. Soft Comput. 118, 108485 (2022)","journal-title":"Appl. Soft Comput."},{"issue":"2","key":"1913_CR42","doi-asserted-by":"publisher","first-page":"1020","DOI":"10.1109\/TCSVT.2023.3289624","volume":"34","author":"X Hu","year":"2024","unstructured":"Hu, X., Zhong, B., Liang, Q., et al.: Transformer tracking via frequency Fusion. IEEE Trans. Circuits Syst. Video Technol. 34(2), 1020\u20131031 (2024)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1913_CR43","doi-asserted-by":"crossref","unstructured":"Peng, B., Alcaide, E., Anthony, Q., et al.: RWKV: reinventing RNNs for the Transformer Era. Find. Assoc. Comput. Linguist. EMNLP 2023, 14048\u201314077 (2023)","DOI":"10.18653\/v1\/2023.findings-emnlp.936"},{"key":"1913_CR44","doi-asserted-by":"crossref","unstructured":"Li, X., Wang, W., Hu, X., et al.: Selective kernel networks[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 510\u2013519 (2019)","DOI":"10.1109\/CVPR.2019.00060"},{"key":"1913_CR45","doi-asserted-by":"crossref","unstructured":"Muller, M., Bibi, A., Giancola, S., et al.: Trackingnet: A large-scale dataset and benchmark for object tracking in the wild[C]\/\/Proceedings of the European Conference on Computer Vision (ECCV). pp. 300\u2013317 (2018)","DOI":"10.1007\/978-3-030-01246-5_19"},{"key":"1913_CR46","doi-asserted-by":"crossref","unstructured":"Fan, H., Lin, L., Yang, F., et al.: Lasot: A high-quality benchmark for large-scale single object tracking[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 5374\u20135383 (2019)","DOI":"10.1109\/CVPR.2019.00552"},{"key":"1913_CR47","doi-asserted-by":"publisher","unstructured":"Huang, L., Zhao, X., Got, K.H.: 10k: A large high-diversity benchmark for generic object tracking in the wild, 43, 1562\u20131577 (2019). https:\/\/doi.org\/10.1109\/TPAMI","DOI":"10.1109\/TPAMI"},{"key":"1913_CR48","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., et al.: Microsoft coco: common objects in context[C]\/\/Computer Vision\u2013ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6\u201312, 2014, Proceedings, Part V 13, pp. 740\u2013755. Springer International Publishing (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"1913_CR49","doi-asserted-by":"crossref","unstructured":"Blatter, P., Kanakis, M., Danelljan, M., et al.: Efficient visual tracking with exemplar transformers[C]\/\/Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision. pp. 1571\u20131581 (2023)","DOI":"10.1109\/WACV56688.2023.00162"},{"key":"1913_CR50","doi-asserted-by":"crossref","unstructured":"Chen, X., Kang, B., Wang, D., et al.: Efficient visual tracking via hierarchical cross-attention transformer[C]\/\/European Conference on Computer Vision, pp. 461\u2013477. Springer Nature Switzerland, Cham (2022)","DOI":"10.1007\/978-3-031-25085-9_26"},{"key":"1913_CR51","doi-asserted-by":"crossref","unstructured":"Cao, Z., Fu, C., Ye, J., et al.: Hift: Hierarchical feature transformer for aerial tracking[C]\/\/Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 15457\u201315466 (2021)","DOI":"10.1109\/ICCV48922.2021.01517"},{"key":"1913_CR52","doi-asserted-by":"crossref","unstructured":"Wei, X., Bai, Y., Zheng, Y., et al.: Autoregressive visual tracking[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp.  9697\u20139706 (2023)","DOI":"10.1109\/CVPR52729.2023.00935"},{"key":"1913_CR53","doi-asserted-by":"crossref","unstructured":"Gao, S., Zhou, C., Zhang, J.: Generalized relation modeling for transformer tracking[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 18686\u201318695 (2023)","DOI":"10.1109\/CVPR52729.2023.01792"},{"key":"1913_CR54","doi-asserted-by":"crossref","unstructured":"Chen, B., Li, P., Bai, L., et al.: Backbone is all your need: a simplified architecture for visual object tracking[C]\/\/European Conference on Computer Vision. Cham: Springer Nature Switzerland. pp. 375\u2013392 (2022)","DOI":"10.1007\/978-3-031-20047-2_22"},{"key":"1913_CR55","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Gool, L.V., Timofte, R.: Probabilistic regression for visual tracking[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 7183\u20137192 (2020)","DOI":"10.1109\/CVPR42600.2020.00721"},{"key":"1913_CR56","doi-asserted-by":"crossref","unstructured":"Li, B., Wu, W., Wang, Q., et al.: Siamrpn++: evolution of siamese visual tracking with very deep networks[C]\/\/Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 4282\u20134291 (2019)","DOI":"10.1109\/CVPR.2019.00441"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01913-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-01913-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01913-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T10:26:25Z","timestamp":1761387985000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-01913-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,21]]},"references-count":56,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["1913"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-01913-3","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2025,8,21]]},"assertion":[{"value":"27 November 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 June 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 August 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"348"}}