{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T00:14:37Z","timestamp":1769645677376,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":18,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819530601","type":"print"},{"value":"9789819530618","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,11,13]],"date-time":"2025-11-13T00:00:00Z","timestamp":1762992000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,13]],"date-time":"2025-11-13T00:00:00Z","timestamp":1762992000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-3061-8_9","type":"book-chapter","created":{"date-parts":[[2025,11,12]],"date-time":"2025-11-12T05:02:26Z","timestamp":1762923746000},"page":"77-84","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["An LLM-Enabled Data Augmentation Framework for\u00a0Low-Resource Scenarios"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8515-2596","authenticated-orcid":false,"given":"Zhongjian","family":"Hu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1184-8117","authenticated-orcid":false,"given":"Peng","family":"Yang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0008-8242-6369","authenticated-orcid":false,"given":"Tianwai","family":"Zhou","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0007-7257-8484","authenticated-orcid":false,"given":"Kun","family":"Song","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,13]]},"reference":[{"issue":"7","key":"9_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3544558","volume":"55","author":"M Bayer","year":"2022","unstructured":"Bayer, M., Kaufhold, M.-A., Reuter, C.: A survey on data augmentation for text classification. ACM Comput. Surv. 55(7), 1\u201339 (2022)","journal-title":"ACM Comput. Surv."},{"key":"9_CR2","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1016\/j.aiopen.2022.03.001","volume":"3","author":"B Li","year":"2022","unstructured":"Li, B., Hou, Y., Che, W.: Data augmentation approaches in natural language processing: a survey. AI Open 3, 71\u201390 (2022)","journal-title":"AI Open"},{"key":"9_CR3","doi-asserted-by":"crossref","unstructured":"\u015eahin, G.G.: To augment or not to augment? A comparative study on text augmentation techniques for low-resource NLP. Comput. Linguist. 48(1), 5\u201342 (2022)","DOI":"10.1162\/coli_a_00425"},{"key":"9_CR4","doi-asserted-by":"crossref","unstructured":"Wei, J., Zou, K.: EDA: easy data augmentation techniques for boosting performance on text classification tasks. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 6382\u20136388 (2019)","DOI":"10.18653\/v1\/D19-1670"},{"key":"9_CR5","doi-asserted-by":"crossref","unstructured":"Zhao, J.: ROR: read-over-read for long document machine reading comprehension. In: Findings of the Association for Computational Linguistics: EMNLP 2021, pp. 1862\u20131872 (2021)","DOI":"10.18653\/v1\/2021.findings-emnlp.160"},{"key":"9_CR6","doi-asserted-by":"crossref","unstructured":"Kumar, V., Choudhary, A., Cho, E.: Data augmentation using pre-trained transformer models. In: Proceedings of the 2nd Workshop on Life-long Learning for Spoken Language Systems, pp. 18\u201326 (2020)","DOI":"10.18653\/v1\/2020.lifelongnlp-1.3"},{"key":"9_CR7","unstructured":"M\u00f8ller, A.G., Dalsgaard, J.A., Pera, A., Aiello, L.M.: Is a prompt and a few samples all you need? Using GPT-4 for data augmentation in low-resource classification tasks. arXiv preprint arXiv:2304.13861 (2023)"},{"key":"9_CR8","unstructured":"Wang, Z., et al.: A comprehensive survey on data augmentation. arXiv preprint arXiv:2405.09591 (2024)"},{"key":"9_CR9","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1007\/978-3-030-22747-0_7","volume-title":"Computational Science \u2013 ICCS 2019","author":"X Wu","year":"2019","unstructured":"Wu, X., Lv, S., Zang, L., Han, J., Hu, S.: Conditional BERT contextual augmentation. In: Rodrigues, J.M.F., et al. (eds.) ICCS 2019, Part IV. LNCS, vol. 11539, pp. 84\u201395. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-22747-0_7"},{"key":"9_CR10","unstructured":"Touvron, H., et al.: Llama: open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)"},{"key":"9_CR11","unstructured":"Team GLM, et\u00a0al.: ChatGLM: a family of large language models from GLM-130B TO GLM-4 all tools. arXiv preprint arXiv:2406.12793 (2024)"},{"key":"9_CR12","unstructured":"Bai, J., et\u00a0al.: Qwen technical report. arXiv preprint arXiv:2309.16609 (2023)"},{"key":"9_CR13","doi-asserted-by":"crossref","unstructured":"Socher, R., et al.: Recursive deep models for semantic compositionality over a sentiment treebank. In: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pp. 1631\u20131642 (2013)","DOI":"10.18653\/v1\/D13-1170"},{"key":"9_CR14","unstructured":"Coucke, A., et al.: Snips voice platform: an embedded spoken language understanding system for private-by-design voice interfaces. arXiv preprint arXiv:1805.10190 (2018)"},{"key":"9_CR15","doi-asserted-by":"crossref","unstructured":"Li, X., Roth, D.: Learning question classifiers. In: COLING 2002: The 19th International Conference on Computational Linguistics (2002)","DOI":"10.3115\/1072228.1072378"},{"key":"9_CR16","doi-asserted-by":"crossref","unstructured":"Ng, N., Yee, K., Baevski, A., Ott, M., Auli, M., Edunov, S.: Facebook fair\u2019s WMT19 news translation task submission. In: Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1), pp. 314\u2013319 (2019)","DOI":"10.18653\/v1\/W19-5333"},{"key":"9_CR17","doi-asserted-by":"crossref","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, volume 1 (long and short papers), pp. 4171\u20134186 (2019)","DOI":"10.18653\/v1\/N19-1423"},{"key":"9_CR18","doi-asserted-by":"crossref","unstructured":"Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7871\u20137880 (2020)","DOI":"10.18653\/v1\/2020.acl-main.703"}],"container-title":["Lecture Notes in Computer Science","Knowledge Science, Engineering and Management"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-3061-8_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,28]],"date-time":"2026-01-28T12:13:55Z","timestamp":1769602435000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-3061-8_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,13]]},"ISBN":["9789819530601","9789819530618"],"references-count":18,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-3061-8_9","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,13]]},"assertion":[{"value":"13 November 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"KSEM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Knowledge Science, Engineering and Management","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Macao","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 August 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 August 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ksem2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ksem2025.scimeeting.cn\/en\/web\/index\/27434","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}