{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T02:38:15Z","timestamp":1769827095785,"version":"3.49.0"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,11]],"date-time":"2025-12-11T00:00:00Z","timestamp":1765411200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,12,11]],"date-time":"2025-12-11T00:00:00Z","timestamp":1765411200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Mach Learn"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s10994-025-06926-7","type":"journal-article","created":{"date-parts":[[2025,12,11]],"date-time":"2025-12-11T22:48:40Z","timestamp":1765493320000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DiverSeed: Integrating Active Learning for Target Domain Data Generation in Instruction Tuning"],"prefix":"10.1007","volume":"115","author":[{"given":"Jingsheng","family":"Gao","sequence":"first","affiliation":[]},{"given":"Mengnan","family":"Qi","sequence":"additional","affiliation":[]},{"given":"Xian","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Suncheng","family":"Xiang","sequence":"additional","affiliation":[]},{"given":"Ke","family":"Ji","sequence":"additional","affiliation":[]},{"given":"Jiacheng","family":"Ruan","sequence":"additional","affiliation":[]},{"given":"Ting","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Yuzhuo","family":"Fu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,11]]},"reference":[{"key":"6926_CR1","unstructured":"Abe, N., Mamitsuka, H. (1998). Query learning strategies using boosting and bagging. In: Proceedings of the Fifteenth International Conference on Machine Learning, pp. 1\u20139."},{"key":"6926_CR2","unstructured":"Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F.L., Almeida, D., Altenschmidt, J., Altman, S., Anadkat, S., et al. (2023). Gpt-4 technical report. arXiv preprint arXiv:2303.08774."},{"key":"6926_CR3","first-page":"23716","volume":"35","author":"J-B Alayrac","year":"2022","unstructured":"Alayrac, J.-B., Donahue, J., Luc, P., Miech, A., Barr, I., Hasson, Y., Lenc, K., Mensch, A., Millican, K., Reynolds, M., et al. (2022). Flamingo: A visual language model for few-shot learning. Advances in Neural Information Processing Systems, 35, 23716\u201323736.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"6926_CR4","doi-asserted-by":"crossref","unstructured":"Bach, S.H., Sanh, V., Yong, Z.-X., Webson, A., Raffel, C., Nayak, N.V., Sharma, A., Kim, T., Bari, M.S., Fevry, T., et al. (2022). Promptsource: An integrated development environment and repository for natural language prompts. arXiv preprint arXiv:2202.01279.","DOI":"10.18653\/v1\/2022.acl-demo.9"},{"key":"6926_CR5","unstructured":"Brinker, K. (2003) Incorporating diversity in active learning with support vector machines. In: Proceedings of the 20th International Conference on Machine Learning (ICML-03), pp. 59\u201366."},{"issue":"1","key":"6926_CR6","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1007\/S10994-023-06454-2","volume":"113","author":"D Cacciarelli","year":"2024","unstructured":"Cacciarelli, D., & Kulahci, M. (2024). Active learning for data streams: A survey. Machine Learning, 113(1), 185\u2013239. https:\/\/doi.org\/10.1007\/S10994-023-06454-2","journal-title":"Machine Learning"},{"key":"6926_CR7","unstructured":"Celikyilmaz, A., Clark, E., Gao, J. (2020). Evaluation of text generation: A survey. arXiv preprint arXiv:2006.14799."},{"key":"6926_CR8","doi-asserted-by":"publisher","unstructured":"Chang, E., Shen, X., Yeh, H.-S., Demberg, V. (2021). On training instance selection for few-shot neural text generation. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pp. 8\u201313. Association for Computational Linguistics, Online . https:\/\/doi.org\/10.18653\/v1\/2021.acl-short.2 . https:\/\/aclanthology.org\/2021.acl-short.2.","DOI":"10.18653\/v1\/2021.acl-short.2"},{"key":"6926_CR9","first-page":"240","volume":"24","author":"A Chowdhery","year":"2023","unstructured":"Chowdhery, A., Narang, S., Devlin, J., Bosma, M., Mishra, G., Roberts, A., Barham, P., Chung, H. W., Sutton, C., Gehrmann, S., Schuh, P., et al. (2023). Palm: Scaling language modeling with pathways. J. Mach. Learn. Res., 24, 240\u20131240113.","journal-title":"J. Mach. Learn. Res."},{"key":"6926_CR10","unstructured":"Chung, H.W., Hou, L., Longpre, S., Zoph, B., Tay, Y., Fedus, W., Li, Y., Wang, X., Dehghani, M., Brahma, S., et al. (2022). Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416."},{"key":"6926_CR11","unstructured":"Clark, P., Cowhey, I., Etzioni, O., Khot, T., Sabharwal, A., Schoenick, C., Tafjord, O. (2018). Think you have solved question answering? try arc, the ai2 reasoning challenge. arXiv preprint arXiv:1803.05457."},{"key":"6926_CR12","unstructured":"Cobbe, K., Kosaraju, V., Bavarian, M., Chen, M., Jun, H., Kaiser, L., Plappert, M., Tworek, J., Hilton, J., Nakano, R., Hesse, C., Schulman, J. (2021). Training verifiers to solve math word problems. CoRR abs\/2110.14168 arxiv2110.14168."},{"key":"6926_CR13","doi-asserted-by":"publisher","unstructured":"Han, C., Du, X., Zhang, C., Lian, Y., Li, X., Gao, M., Wang, B. (2023). DialCoT meets PPO: Decomposing and exploring reasoning paths in smaller language models. In: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pp. 8055\u20138068. Association for Computational Linguistics, Singapore . https:\/\/doi.org\/10.18653\/v1\/2023.emnlp-main.501 . https:\/\/aclanthology.org\/2023.emnlp-main.501.","DOI":"10.18653\/v1\/2023.emnlp-main.501"},{"key":"6926_CR14","doi-asserted-by":"crossref","unstructured":"Hosseini, M.J., Hajishirzi, H., Etzioni, O., Kushman, N. (2014). Learning to solve arithmetic word problems with verb categorization. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 523\u2013533.","DOI":"10.3115\/v1\/D14-1058"},{"key":"6926_CR16","unstructured":"Hu, E.J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., Chen, W. (2021). Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685."},{"key":"6926_CR15","doi-asserted-by":"crossref","unstructured":"Hu, Z., Lan, Y., Wang, L., Xu, W., Lim, E.-P., Lee, R.K.-W., Bing, L., Poria, S. (2023). Llm-adapters: An adapter family for parameter-efficient fine-tuning of large language models. arXiv preprint arXiv:2304.01933.","DOI":"10.18653\/v1\/2023.emnlp-main.319"},{"key":"6926_CR17","doi-asserted-by":"crossref","unstructured":"Ienco, D., Bifet, A., \u017dliobait\u0117, I., Pfahringer, B. (2013). Clustering based active learning for evolving data streams. In: International Conference on Discovery Science, pp. 79\u201393 . Springer.","DOI":"10.1007\/978-3-642-40897-7_6"},{"key":"6926_CR18","unstructured":"Iyer, S., Lin, X.V., Pasunuru, R., Mihaylov, T., Simig, D., Yu, P., Shuster, K., Wang, T., Liu, Q., Koura, P.S., et al. (2022). Opt-iml: Scaling language model instruction meta learning through the lens of generalization. arXiv preprint arXiv:2212.12017."},{"key":"6926_CR19","unstructured":"Jiang, A.Q., Sablayrolles, A., Mensch, A., Bamford, C., Chaplot, D.S., Casas, D.d.l., Bressand, F., Lengyel, G., Lample, G., Saulnier, L., et al. (2023). Mistral 7b. arXiv preprint arXiv:2310.06825."},{"issue":"14","key":"6926_CR20","doi-asserted-by":"publisher","DOI":"10.3390\/app11146421","volume":"11","author":"D Jin","year":"2021","unstructured":"Jin, D., Pan, E., Oufattole, N., Weng, W.-H., Fang, H., & Szolovits, P. (2021). What disease does this patient have? A large-scale open domain question answering dataset from medical exams. Applied Sciences, 11(14), Article 6421.","journal-title":"Applied Sciences"},{"key":"6926_CR21","unstructured":"Kirsch, A., Van\u00a0Amersfoort, J., Gal, Y. (2019). Batchbald: Efficient and diverse batch acquisition for deep bayesian active learning. Advances in neural information processing systems 32."},{"key":"6926_CR22","first-page":"22199","volume":"35","author":"T Kojima","year":"2022","unstructured":"Kojima, T., Gu, S. S., Reid, M., Matsuo, Y., & Iwasawa, Y. (2022). Large language models are zero-shot reasoners. Advances In Neural Information Processing Systems, 35, 22199\u201322213.","journal-title":"Advances In Neural Information Processing Systems"},{"key":"6926_CR23","doi-asserted-by":"publisher","first-page":"585","DOI":"10.1162\/tacl_a_00160","volume":"3","author":"R Koncel-Kedziorski","year":"2015","unstructured":"Koncel-Kedziorski, R., Hajishirzi, H., Sabharwal, A., Etzioni, O., & Ang, S. D. (2015). Parsing algebraic word problems into equations. Transactions of the Association for Computational Linguistics, 3, 585\u2013597.","journal-title":"Transactions of the Association for Computational Linguistics"},{"key":"6926_CR24","unstructured":"Kuhn, L., Gal, Y., Farquhar, S. (2023). Semantic uncertainty: Linguistic invariances for uncertainty estimation in natural language generation. arXiv preprint arXiv:2302.09664."},{"key":"6926_CR25","unstructured":"Lee, N., Sreenivasan, K., Lee, J.D., Lee, K., Papailiopoulos, D. (2023). Teaching arithmetic to small transformers. arXiv preprint arXiv:2307.03381."},{"key":"6926_CR26","doi-asserted-by":"crossref","unstructured":"Lin, Z., Liu, J.Z., Shang, J. (2022). Towards collaborative neural-symbolic graph semantic parsing via uncertainty. Findings of the Association for Computational Linguistics: ACL 2022.","DOI":"10.18653\/v1\/2022.findings-acl.328"},{"key":"6926_CR27","doi-asserted-by":"crossref","unstructured":"Ling, W., Yogatama, D., Dyer, C., Blunsom, P. (2017). Program induction by rationale generation: Learning to solve and explain algebraic word problems. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 158\u2013167.","DOI":"10.18653\/v1\/P17-1015"},{"key":"6926_CR29","doi-asserted-by":"crossref","unstructured":"Liu, A., Swayamdipta, S., Smith, N.A., Choi, Y. (2022). Wanli: Worker and ai collaboration for natural language inference dataset creation. arXiv preprint arXiv:2201.05955.","DOI":"10.18653\/v1\/2022.findings-emnlp.508"},{"key":"6926_CR28","doi-asserted-by":"crossref","unstructured":"Liu, H., Li, C., Li, Y., Lee, Y.J. (2024). Improved baselines with visual instruction tuning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 26296\u201326306.","DOI":"10.1109\/CVPR52733.2024.02484"},{"key":"6926_CR30","unstructured":"Luo, H., Sun, Q., Xu, C., Zhao, P., Lou, J., Tao, C., Geng, X., Lin, Q., Chen, S., Zhang, D. (2023). Wizardmath: Empowering mathematical reasoning for large language models via reinforced evol-instruct. arXiv preprint arXiv:2308.09583."},{"key":"6926_CR31","unstructured":"OpenAI: Introducing chatgpt (2023)."},{"key":"6926_CR32","first-page":"27730","volume":"35","author":"L Ouyang","year":"2022","unstructured":"Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., et al. (2022). Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35, 27730\u201327744.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"6926_CR33","doi-asserted-by":"publisher","unstructured":"Patel, A., Bhattamishra, S., Goyal, N. (2021). Are NLP models really able to solve simple math word problems? In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, Online, June 6-11, 2021, pp. 2080\u20132094. Association for Computational Linguistics, ??? . https:\/\/doi.org\/10.18653\/V1\/2021.NAACL-MAIN.168.","DOI":"10.18653\/V1\/2021.NAACL-MAIN.168"},{"key":"6926_CR34","unstructured":"Peng, B., Li, C., He, P., Galley, M., Gao, J. (2023). Instruction tuning with gpt-4. arXiv preprint arXiv:2304.03277."},{"key":"6926_CR35","doi-asserted-by":"crossref","unstructured":"Reimers, N., Gurevych, I. (2019). Sentence-bert: Sentence embeddings using siamese bert-networks. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 3982\u20133992.","DOI":"10.18653\/v1\/D19-1410"},{"issue":"9","key":"6926_CR36","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3472291","volume":"54","author":"P Ren","year":"2021","unstructured":"Ren, P., Xiao, Y., Chang, X., Huang, P.-Y., Li, Z., Gupta, B. B., Chen, X., & Wang, X. (2021). A survey of deep active learning. Acm Computing Surveys, 54(9), 1\u201340.","journal-title":"Acm Computing Surveys"},{"key":"6926_CR37","doi-asserted-by":"crossref","unstructured":"Roy, S., Roth, D. (2016). Solving general arithmetic word problems. CoRR abs\/1608.01413 arXiv:1608.01413.","DOI":"10.18653\/v1\/D15-1202"},{"key":"6926_CR38","doi-asserted-by":"crossref","unstructured":"Schick, T., Sch\u00fctze, H. (2021). Generating datasets with pretrained language models. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 6943\u20136951.","DOI":"10.18653\/v1\/2021.emnlp-main.555"},{"key":"6926_CR39","unstructured":"Sener, O., Savarese, S. (2018). Active learning for convolutional neural networks: A core-set approach. In: International Conference on Learning Representations."},{"key":"6926_CR40","doi-asserted-by":"publisher","first-page":"164","DOI":"10.1007\/s10618-016-0460-3","volume":"31","author":"M Sharma","year":"2017","unstructured":"Sharma, M., & Bilgic, M. (2017). Evidence-based uncertainty sampling for active learning. Data Mining And Knowledge Discovery, 31, 164\u2013202.","journal-title":"Data Mining And Knowledge Discovery"},{"key":"6926_CR41","unstructured":"Sun, Z., Shen, Y., Zhou, Q., Zhang, H., Chen, Z., Cox, D., Yang, Y., Gan, C. (2023). Principle-driven self-alignment of language models from scratch with minimal human supervision. arXiv preprint arXiv:2305.03047."},{"issue":"4","key":"6926_CR42","doi-asserted-by":"publisher","DOI":"10.3390\/math11040820","volume":"11","author":"A Tharwat","year":"2023","unstructured":"Tharwat, A., & Schenck, W. (2023). A survey on active learning: State-of-the-art, practical challenges and research directions. Mathematics, 11(4), Article 820.","journal-title":"Mathematics"},{"key":"6926_CR43","unstructured":"Thoppilan, R., De\u00a0Freitas, D., Hall, J., Shazeer, N., Kulshreshtha, A., Cheng, H., Jin, A., Bos, T., Baker, L., Du, Y., et al. (2022). Lamda: Language models for dialog applications. CoRR abs\/2201.08239 arXiv:2201.08239."},{"key":"6926_CR44","unstructured":"Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., et al. (2023). Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288."},{"key":"6926_CR46","doi-asserted-by":"publisher","unstructured":"Wang, S., Liu, Y., Xu, Y., Zhu, C., Zeng, M. (2021). Want to reduce labeling cost? GPT-3 can help. In: Findings of the Association for Computational Linguistics: EMNLP 2021, pp. 4195\u20134205. Association for Computational Linguistics, Punta Cana, Dominican Republic . https:\/\/doi.org\/10.18653\/v1\/2021.findings-emnlp.354. https:\/\/aclanthology.org\/2021.findings-emnlp.354.","DOI":"10.18653\/v1\/2021.findings-emnlp.354"},{"key":"6926_CR47","unstructured":"Wang, Z., Yu, A.W., Firat, O., Cao, Y. (2021). Towards zero-label language learning. arXiv preprint arXiv:2109.09193."},{"key":"6926_CR45","doi-asserted-by":"publisher","unstructured":"Wang, Y., Kordi, Y., Mishra, S., Liu, A., Smith, N.A., Khashabi, D., Hajishirzi, H. (2023). Self-instruct: Aligning language models with self-generated instructions. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, Toronto, Canada, July 9-14, 2023, pp. 13484\u201313508. Association for Computational Linguistics, ??? . https:\/\/doi.org\/10.18653\/V1\/2023.ACL-LONG.754.","DOI":"10.18653\/V1\/2023.ACL-LONG.754"},{"key":"6926_CR48","unstructured":"Wei, J., Bosma, M., Zhao, V.Y., Guu, K., Yu, A.W., Lester, B., Du, N., Dai, A.M., Le, Q.V. (2022). Finetuned language models are zero-shot learners. In: The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net, ??? . https:\/\/openreview.net\/forum?id=gEZrGCozdqR."},{"key":"6926_CR49","doi-asserted-by":"crossref","unstructured":"Wu, Y., Kozintsev, I., Bouguet, J.-Y., Dulong, C. (2006). Sampling strategies for active learning in personal photo retrieval. In: 2006 IEEE International Conference on Multimedia and Expo, pp. 529\u2013532 . IEEE.","DOI":"10.1109\/ICME.2006.262442"},{"key":"6926_CR50","doi-asserted-by":"publisher","unstructured":"Xu, C., Guo, D., Duan, N., McAuley, J.J. (2023). Baize: An open-source chat model with parameter-efficient tuning on self-chat data. CoRR abs\/2304.01196 https:\/\/doi.org\/10.48550\/ARXIV.2304.01196.","DOI":"10.48550\/ARXIV.2304.01196"},{"key":"6926_CR51","unstructured":"Yang, A., Yang, B., Hui, B., Zheng, B., Yu, B., Zhou, C., Li, C., Li, C., Liu, D., Huang, F., et al. (2024). Qwen2 technical report. arXiv preprint arXiv:2407.10671"},{"key":"6926_CR52","doi-asserted-by":"publisher","unstructured":"Ye, J., Gao, J., Li, Q., Xu, H., Feng, J., Wu, Z., Yu, T., Kong, L. (2022). ZeroGen: Efficient zero-shot learning via dataset generation. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 11653\u201311669. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates . https:\/\/doi.org\/10.18653\/v1\/2022.emnlp-main.801 . https:\/\/aclanthology.org\/2022.emnlp-main.801.","DOI":"10.18653\/v1\/2022.emnlp-main.801"},{"key":"6926_CR53","doi-asserted-by":"crossref","unstructured":"Yoo, D., Kweon, I.S. (2019). Learning loss for active learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 93\u2013102.","DOI":"10.1109\/CVPR.2019.00018"},{"key":"6926_CR54","unstructured":"Yu, L., Jiang, W., Shi, H., Yu, J., Liu, Z., Zhang, Y., Kwok, J.T., Li, Z., Weller, A., Liu, W. (2023). Metamath: Bootstrap your own mathematical questions for large language models. arXiv preprint arXiv:2309.12284."},{"key":"6926_CR55","unstructured":"Yu, Y., Zhuang, Y., Zhang, J., Meng, Y., Ratner, A., Krishna, R., Shen, J., Zhang, C. (2023). Large language model as attributed training data generator: A tale of diversity and bias. arXiv preprint arXiv:2306.15895."},{"key":"6926_CR56","unstructured":"Zhang, Z., Zhang, A., Li, M., Smola, A. (2022). Automatic chain of thought prompting in large language models. arXiv preprint arXiv:2210.03493."},{"key":"6926_CR57","doi-asserted-by":"crossref","unstructured":"Zhong, R., Lee, K., Zhang, Z., Klein, D. (2021). Adapting language models for zero-shot learning by meta-tuning on dataset and prompt collections. In: Findings of the Association for Computational Linguistics: EMNLP 2021, pp. 2856\u20132878.","DOI":"10.18653\/v1\/2021.findings-emnlp.244"},{"key":"6926_CR58","unstructured":"Zhou, C., Liu, P., Xu, P., Iyer, S., Sun, J., Mao, Y., Ma, X., Efrat, A., Yu, P., Yu, L., et al. (2023). Lima: Less is more for alignment. arXiv preprint arXiv:2305.11206."}],"container-title":["Machine Learning"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10994-025-06926-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10994-025-06926-7","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10994-025-06926-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T14:05:16Z","timestamp":1769781916000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10994-025-06926-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,11]]},"references-count":58,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["6926"],"URL":"https:\/\/doi.org\/10.1007\/s10994-025-06926-7","relation":{},"ISSN":["0885-6125","1573-0565"],"issn-type":[{"value":"0885-6125","type":"print"},{"value":"1573-0565","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,11]]},"assertion":[{"value":"22 January 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 July 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 October 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 December 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}}],"article-number":"1"}}