{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,5]],"date-time":"2026-05-05T12:01:22Z","timestamp":1777982482003,"version":"3.51.4"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T00:00:00Z","timestamp":1770076800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T00:00:00Z","timestamp":1770076800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the Key Project of the National Language Commission","award":["ZDI145-110"],"award-info":[{"award-number":["ZDI145-110"]}]},{"name":"Scientific research project of the Ministry of Education","award":["25C11417008"],"award-info":[{"award-number":["25C11417008"]}]},{"name":"the Key Laboratory Project","award":["YYZN-2024-6"],"award-info":[{"award-number":["YYZN-2024-6"]}]},{"name":"the China Disabled Persons' Federation Project","award":["2024CDPFAT-22"],"award-info":[{"award-number":["2024CDPFAT-22"]}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62171042, U24A20331"],"award-info":[{"award-number":["62171042, U24A20331"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"the R&D Program of the Beijing Municipal Education Commission","award":["KZ202211417048"],"award-info":[{"award-number":["KZ202211417048"]}]},{"name":"the Project for the Construction and Support of High-Level Innovative Teams in Beijing Municipal Institutions","award":["BPHR20220121"],"award-info":[{"award-number":["BPHR20220121"]}]},{"name":"the Beijing Natural Science Foundation","award":["4232026, 4242020"],"award-info":[{"award-number":["4232026, 4242020"]}]},{"name":"the Academic Research Projects of Beijing Union University","award":["ZK20202514"],"award-info":[{"award-number":["ZK20202514"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1007\/s00530-025-02153-1","type":"journal-article","created":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T03:47:41Z","timestamp":1770090461000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["TSPPO: transformer-based sequential proximal policy optimization for multi-agent systems"],"prefix":"10.1007","volume":"32","author":[{"given":"Tao","family":"Yang","sequence":"first","affiliation":[]},{"given":"Yuxiao","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Cheng","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Hongzhe","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,3]]},"reference":[{"key":"2153_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.121111","volume":"235","author":"F Ren","year":"2024","unstructured":"Ren, F., Dong, W., Zhao, X., Zhang, F., Kong, Y., Yang, Q.: Two-layer coordinated reinforcement learning for traffic signal control in traffic network. Expert Syst. Appl. 235, 121111 (2024)","journal-title":"Expert Syst. Appl."},{"issue":"2","key":"2153_CR2","doi-asserted-by":"publisher","first-page":"1071","DOI":"10.1109\/TIV.2024.3422506","volume":"10","author":"X Zhang","year":"2025","unstructured":"Zhang, X., Zhang, H., Tang, H., Liang, L., Cheng, L., Chen, X., Ding, W., Zhang, X.-P.: A scalable mean-field marl framework for multi-objective v2x resource allocation. IEEE Trans. Intell. Veh. 10(2), 1071\u20131086 (2025)","journal-title":"IEEE Trans. Intell. Veh."},{"key":"2153_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2022.103514","volume":"126","author":"SSA Zaidi","year":"2022","unstructured":"Zaidi, S.S.A., Ansari, M.S., Aslam, A., Kanwal, N., Asghar, M., Lee, B.: A survey of modern deep learning based object detection models. Digital Signal Process. 126, 103514 (2022)","journal-title":"Digital Signal Process."},{"issue":"5","key":"2153_CR4","doi-asserted-by":"publisher","first-page":"2991","DOI":"10.1007\/s00530-022-00922-w","volume":"29","author":"W Chen","year":"2023","unstructured":"Chen, W., Nie, J.: A maddpg-based multi-agent antagonistic algorithm for sea battlefield confrontation. Multimed. Syst. 29(5), 2991\u20133000 (2023)","journal-title":"Multimed. Syst."},{"issue":"3","key":"2153_CR5","doi-asserted-by":"publisher","first-page":"4141","DOI":"10.1007\/s40747-024-01389-0","volume":"10","author":"E Zhao","year":"2024","unstructured":"Zhao, E., Zhou, N., Liu, C., Su, H., Liu, Y., Cong, J.: Time-aware maddpg with lstm for multi-agent obstacle avoidance: a comparative study. Complex Intell. Syst. 10(3), 4141\u20134155 (2024)","journal-title":"Complex Intell. Syst."},{"key":"2153_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2024.124959","volume":"257","author":"T Cui","year":"2024","unstructured":"Cui, T., Yang, X., Jia, F., Jin, J., Ye, Y., Bai, R.: Mobile robot sequential decision making using a deep reinforcement learning hyper-heuristic approach. Expert Syst. Appl. 257, 124959 (2024)","journal-title":"Expert Syst. Appl."},{"key":"2153_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2024.120119","volume":"661","author":"Y Zhou","year":"2024","unstructured":"Zhou, Y., Tao, Y., Lei, X., Peng, X.: Self-organized swarm robot for multi-target trapping based on self-regulated density interaction. Inf. Sci. 661, 120119 (2024)","journal-title":"Inf. Sci."},{"key":"2153_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2022.103489","volume":"126","author":"N Xing","year":"2022","unstructured":"Xing, N., Wang, Y., Teng, L., Li, L.: Communication and trajectory design in uav-enabled flying network. Digit. Signal Process 126, 103489 (2022)","journal-title":"Digit. Signal Process"},{"issue":"7","key":"2153_CR9","first-page":"96","volume":"12","author":"T Yang","year":"2025","unstructured":"Yang, T., Shi, X., Xu, C., Yang, Y., Liu, H., Zeng, Q.: Sd2ac: a reinforcement learning framework using distribution evaluation and sequential decision-making for ucav combat. J. Comput. Des. Eng. 12(7), 96\u2013112 (2025)","journal-title":"J. Comput. Des. Eng."},{"key":"2153_CR10","doi-asserted-by":"publisher","first-page":"5877","DOI":"10.1109\/TASE.2024.3432405","volume":"22","author":"Y Wang","year":"2025","unstructured":"Wang, Y., Boyle, D.: Constrained reinforcement learning using distributional representation for trustworthy quadrotor uav tracking control. IEEE Trans. Autom. Sci. Eng. 22, 5877\u20135894 (2025)","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"2153_CR11","doi-asserted-by":"crossref","unstructured":"Zhao Y, Ju L, Hern\u00e1ndez-Orallo J. Team formation through an assessor: choosing MARL agents in pursuit\u2013evasion games[J]. Complex & Intelligent Systems.10(3), 3473-3492(2024)","DOI":"10.1007\/s40747-023-01336-5"},{"key":"2153_CR12","unstructured":"Pasukonis, J., Lillicrap, T., Hafner, D.: Evaluating long-term memory in 3d mazes, arXiv preprint arXiv:2210.13383 (2022)"},{"key":"2153_CR13","unstructured":"Osband, I., Doron, Y., Hessel, M., Aslanides, J., Sezener, E., Saraiva, A., McKinney, K., Lattimore, T., Szepesv\u00e1ri, C., Singh, S., Van Roy, B., Sutton, R., Silver, D., van Hasselt, H.: Behaviour suite for reinforcement learning. In: International Conference on Learning Representations, pp. 1010\u20131023, (2020)"},{"key":"2153_CR14","unstructured":"Parisotto, E., Song, F., Rae, J., Pascanu, R., Gulcehre, C., Jayakumar, S., Jaderberg, M., Kaufman, R.L., Clark, A., Noury, S. et\u00a0al.: Stabilizing transformers for reinforcement learning. In: International conference on machine learning, PMLR, pp. 7487\u20137498, (2020)"},{"issue":"4","key":"2153_CR15","doi-asserted-by":"publisher","first-page":"2181","DOI":"10.1007\/s00530-023-01091-0","volume":"29","author":"Y Zhao","year":"2023","unstructured":"Zhao, Y., Tan, Z., Bao, B.-K., Tu, Z.: Centralized sub-critic based hierarchical-structured reinforcement learning for temporal sentence grounding. Multimed. Syst. 29(4), 2181\u20132191 (2023)","journal-title":"Multimed. Syst."},{"issue":"5","key":"2153_CR16","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3453160","volume":"54","author":"S Pateria","year":"2021","unstructured":"Pateria, S., Subagdja, B., Tan, A.-H., Quek, C.: Hierarchical reinforcement learning: a comprehensive survey. ACM Comput. Surv. (CSUR) 54(5), 1\u201335 (2021)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"2153_CR17","doi-asserted-by":"publisher","first-page":"359","DOI":"10.1016\/j.neunet.2023.02.037","volume":"162","author":"J Wang","year":"2023","unstructured":"Wang, J., Yuan, M., Li, Y., Zhao, Z.: Hierarchical attention master-slave for heterogeneous multi-agent reinforcement learning. Neural Netw. 162, 359\u2013368 (2023)","journal-title":"Neural Netw."},{"key":"2153_CR18","doi-asserted-by":"crossref","unstructured":"Liu, Z., Xu, Z., Fan, G.: Hierarchical multi-agent reinforcement learning with intrinsic reward rectification. In: ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1\u20135, (2023)","DOI":"10.1109\/ICASSP49357.2023.10095374"},{"issue":"1","key":"2153_CR19","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1038\/s42256-021-00433-9","volume":"4","author":"M Eppe","year":"2022","unstructured":"Eppe, M., Gumbsch, C., Kerzel, M., Nguyen, P.D., Butz, M.V., Wermter, S.: Intelligent problem-solving as integrated hierarchical reinforcement learning. Nat. Mach. Intell. 4(1), 11\u201320 (2022)","journal-title":"Nat. Mach. Intell."},{"key":"2153_CR20","doi-asserted-by":"crossref","unstructured":"Yu X, Lin Y, Wang X, et al. GHQ: grouped hybrid Q-learning for cooperative heterogeneous multi-agent reinforcement learning[J]. Complex & Intelligent Systems.10(4), 5261-5280(2024)","DOI":"10.1007\/s40747-024-01415-1"},{"issue":"10","key":"2153_CR21","first-page":"11735","volume":"37","author":"Z Xu","year":"2023","unstructured":"Xu, Z., Bai, Y., Zhang, B., Li, D., Fan, G.: Haven: hierarchical cooperative multi-agent reinforcement learning with dual coordination mechanism. Proc. AAAI Conf. Artif. Intell. 37(10), 11735\u201311743 (2023)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"2153_CR22","first-page":"61759","volume-title":"Advances in Neural Information Processing Systems","author":"M Yang","year":"2023","unstructured":"Yang, M., Yang, Y., Lu, Z., Zhou, W., Li, H.: Hierarchical multi-agent skill discovery. In: Oh, A., Naumann, T., Globerson, A., Saenko, K., Hardt, M., Levine, S. (eds.) Advances in Neural Information Processing Systems, vol. 36, pp. 61759\u201361776. Curran Associates Inc (2023)"},{"key":"2153_CR23","unstructured":"Kuba, J.G., Wen, M., Meng, L., gu, s., Zhang, H., Mguni, D., Wang, J., Yang, Y.: Settling the variance of multi-agent policy gradients, in: M.\u00a0Ranzato, A.\u00a0Beygelzimer, Y.\u00a0Dauphin, P.\u00a0Liang, J.\u00a0W. Vaughan (Eds.), Advances in Neural Information Processing Systems, Vol.\u00a034, Curran Associates, Inc., pp. 13458\u201313470, (2021)"},{"key":"2153_CR24","volume-title":"Trust region policy optimisation in multi-agent reinforcement learning","author":"J Kuba","year":"2021","unstructured":"Kuba, J., Chen, R., Wen, M., Wen, Y., Sun, F., Wang, J., Yang, Y.: Trust region policy optimisation in multi-agent reinforcement learning. Cornell University - arXiv, Cornell University - arXiv (Sep (2021)"},{"key":"2153_CR25","unstructured":"Wang, X., Tian, Z., Wan, Z., Wen, Y., Wang, J., Zhang, W.: Order matters: Agent-by-agent policy optimization, arXiv preprint arXiv:2302.06205 (2023)"},{"key":"2153_CR26","unstructured":"Liu, J., Zhong, Y., Hu, S., Fu, H., Fu, Q., Chang, X., Yang, Y.: Maximum entropy heterogeneous-agent mirror learning, arXiv preprint arXiv:2306.10715 (2023)"},{"key":"2153_CR27","first-page":"8536","volume":"37","author":"C Li","year":"2023","unstructured":"Li, C., Liu, J., Zhang, Y., Wei, Y., Niu, Y., Yang, Y., Liu, Y., Ouyang, W.: Ace: cooperative multi-agent q-learning with bidirectional action-dependency. Proc. AAAI Conf. Artif. Intell. 37, 8536\u20138544 (2023)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"2153_CR28","first-page":"16509","volume":"35","author":"M Wen","year":"2022","unstructured":"Wen, M., Kuba, J., Lin, R., Zhang, W., Wen, Y., Wang, J., Yang, Y.: Multi-agent reinforcement learning is a sequence modeling problem. Adv. Neural. Inf. Process. Syst. 35, 16509\u201316521 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2153_CR29","unstructured":"Kuba, J.G., Feng, X., Ding, S., Dong, H., Wang, J., Yang, Y.: Heterogeneous-agent mirror learning: a continuum of solutions to cooperative marl, arXiv preprint arXiv:2208.01682 (2022)"},{"key":"2153_CR30","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A., Kaiser, L., Polosukhin, I.: Attention is all you need. Neural Information Processing Systems, Neural Information Processing Systems (2017)"},{"issue":"4","key":"2153_CR31","doi-asserted-by":"publisher","first-page":"819","DOI":"10.1287\/moor.27.4.819.297","volume":"27","author":"DS Bernstein","year":"2002","unstructured":"Bernstein, D.S., Givan, R., Immerman, N., Zilberstein, S.: The complexity of decentralized control of markov decision processes. Math. Oper. Res. 27(4), 819\u2013840 (2002)","journal-title":"Math. Oper. Res."},{"key":"2153_CR32","doi-asserted-by":"crossref","unstructured":"Garcia, F., Rachelson, E.: Markov decision processes, Markov Decision Processes in Artificial Intelligence 1\u201338 (2013)","DOI":"10.1002\/9781118557426.ch1"},{"key":"2153_CR33","doi-asserted-by":"crossref","unstructured":"Nash, J.: Non-cooperative games, Annals of mathematics, 286\u2013295 (1951)","DOI":"10.2307\/1969529"},{"key":"2153_CR34","first-page":"1273","volume":"34","author":"M Janner","year":"2021","unstructured":"Janner, M., Li, Q., Levine, S.: Offline reinforcement learning as one big sequence modeling problem. Adv. Neural. Inf. Process. Syst. 34, 1273\u20131286 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2153_CR35","first-page":"27921","volume":"35","author":"K-H Lee","year":"2022","unstructured":"Lee, K.-H., Nachum, O., Yang, M.S., Lee, L., Freeman, D., Guadarrama, S., Fischer, I., Xu, W., Jang, E., Michalewski, H., et al.: Multi-game decision transformers. Adv. Neural. Inf. Process. Syst. 35, 27921\u201327936 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2153_CR36","unstructured":"Davis, J.Q., Gu, A., Choromanski, K., Dao, T., Re, C., Finn, C., Liang, P.: Catformer: designing stable transformers via sensitivity analysis. In: International Conference on Machine Learning, PMLR, pp. 2489\u20132499, (2021)"},{"key":"2153_CR37","unstructured":"Banino, A., Badia, A.P., Walker, J., Scholtes, T., Mitrovic, J., Blundell, C.: Coberl: Contrastive bert for reinforcement learning, arXiv preprint arXiv:2107.05431 (2021)"},{"key":"2153_CR38","unstructured":"Stooke, A., Lee, K., Abbeel, P., Laskin, M.: Decoupling representation learning from reinforcement learning. In: International Conference on Machine Learning, PMLR, pp. 9870\u20139879, (2021)"},{"key":"2153_CR39","unstructured":"Schulman, J., Levine, S., Abbeel, P., Jordan, M., Moritz, P.: Trust region policy optimization. In: International conference on machine learning, PMLR, pp. 1889\u20131897, (2015)"},{"key":"2153_CR40","unstructured":"Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy optimization algorithms, arXiv preprint arXiv:1707.06347 (2017)"},{"key":"2153_CR41","first-page":"24611","volume":"35","author":"C Yu","year":"2022","unstructured":"Yu, C., Velu, A., Vinitsky, E., Gao, J., Wang, Y., Bayen, A., Wu, Y.: The surprising effectiveness of ppo in cooperative multi-agent games. Adv. Neural. Inf. Process. Syst. 35, 24611\u201324624 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2153_CR42","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need, in: Advances in Neural Information Processing Systems, Vol.\u00a030, Curran Associates, Inc., Long Beach, CA, USA, pp. 5998\u20136008, (2017)"},{"key":"2153_CR43","doi-asserted-by":"crossref","unstructured":"Cu\u00e9llar, M.P., Delgado, M., Pegalajar, M.: An application of non-linear programming to train recurrent neural networks in time series prediction problems, in: Enterprise information systems VII, Springer, pp. 95\u2013102 (2006)","DOI":"10.1007\/978-1-4020-5347-4_11"},{"key":"2153_CR44","unstructured":"Gupta, P., Runkler, T., Adel, H., Andrassy, B., Zimmermann, H.-G., Sch\u00fctze, H.: Deep learning methods for the extraction of relations in natural language text, Master\u2019s thesis, Technical University of Munich, Germany (2015)"},{"key":"2153_CR45","doi-asserted-by":"crossref","unstructured":"Li, J., Kuang, K., Wang, B., Liu, F., Chen, L., Wu, F., Xiao, J.: Shapley counterfactual credits for multi-agent reinforcement learning. In: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 934\u2013942, (2021)","DOI":"10.1145\/3447548.3467420"},{"key":"2153_CR46","doi-asserted-by":"crossref","unstructured":"Winter, E.: The shapley value. Handbook of game theory with economic applications 3, 2025\u20132054 (2002)","DOI":"10.1016\/S1574-0005(02)03016-3"},{"issue":"1","key":"2153_CR47","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/s10107-015-0892-3","volume":"151","author":"SJ Wright","year":"2015","unstructured":"Wright, S.J.: Coordinate descent algorithms. Math. Program. 151(1), 3\u201334 (2015)","journal-title":"Math. Program."},{"key":"2153_CR48","unstructured":"Rashid, T., Samvelyan, M., Witt, C., Farquhar, G., Foerster, J., Whiteson, S.: Qmix: monotonic value function factorisation for deep multi-agent reinforcement learning, arXiv: Learning,arXiv: Learning (2018)"},{"key":"2153_CR49","unstructured":"Lowe R, Wu Y I, Tamar A, et al. Multi-agent actor-critic for mixed cooperative-competitive environments[J]. Advances in neural information processing systems.30,1-12(2017)"},{"key":"2153_CR50","unstructured":"Ackermann, J., Gabler, V., Osa, T., Sugiyama, M.: Reducing overestimation bias in multi-agent domains using double centralized critics, arXiv preprint arXiv:1910.01465 (2019)"},{"key":"2153_CR51","doi-asserted-by":"crossref","unstructured":"Sunehag, P., Lever, G., Gruslys, A., Czarnecki, W.M., Zambaldi, V., Jaderberg, M., Lanctot, M., Sonnerat, N., Leibo, J.Z., Tuyls, K. et\u00a0al.: Value-decomposition networks for cooperative multi-agent learning, arXiv preprint arXiv:1706.05296 (2017)","DOI":"10.65109\/JSRC7365"},{"key":"2153_CR52","unstructured":"Fu, W., Yu, C., Xu, Z., Yang, J., Wu, Y.: Revisiting some common practices in cooperative multi-agent reinforcement learning. In: International Conference on Machine Learning, PMLR, pp. 6863\u20136877, (2022)"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02153-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-02153-1","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02153-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,18]],"date-time":"2026-04-18T04:28:19Z","timestamp":1776486499000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-02153-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,3]]},"references-count":52,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,4]]}},"alternative-id":["2153"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-02153-1","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-6780777\/v1","asserted-by":"object"}]},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,3]]},"assertion":[{"value":"30 May 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 February 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The research conducted for this study adheres to ethical standards. The data utilized in this study are sourced from previously published literature or publicly available resources, ensuring compliance with ethical norms in research.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical and informed consent for data used"}}],"article-number":"118"}}