{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T19:24:12Z","timestamp":1757618652933,"version":"3.44.0"},"reference-count":32,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,7,9]],"date-time":"2025-07-09T00:00:00Z","timestamp":1752019200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,7,9]],"date-time":"2025-07-09T00:00:00Z","timestamp":1752019200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SN COMPUT. SCI."],"DOI":"10.1007\/s42979-025-04120-z","type":"journal-article","created":{"date-parts":[[2025,7,10]],"date-time":"2025-07-10T09:45:33Z","timestamp":1752140733000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Text Augmentation for Enhancing the Text Classification for Low Resource Language"],"prefix":"10.1007","volume":"6","author":[{"given":"Krish","family":"Kumar","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7651-3820","authenticated-orcid":false,"given":"Bhawana","family":"Rudra","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,9]]},"reference":[{"key":"4120_CR1","doi-asserted-by":"publisher","first-page":"842","DOI":"10.1162\/tacl_a_00349","volume":"8","author":"A Rogers","year":"2021","unstructured":"Rogers A, Kovaleva O, Rumshisky A. A primer in Bertology: what we know about how Bert works. Trans Association Comput Linguistics. 2021;8:842\u201366.","journal-title":"Trans Association Comput Linguistics"},{"key":"4120_CR2","doi-asserted-by":"crossref","unstructured":"Erk K, Pad\u00b4o S. A structured vector space model for word meaning in con- text. In: Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing, pp. 897\u2013906 (2008).","DOI":"10.3115\/1613715.1613831"},{"key":"4120_CR3","doi-asserted-by":"publisher","first-page":"1191","DOI":"10.1007\/s10579-013-9233-4","volume":"47","author":"L Borin","year":"2013","unstructured":"Borin L, Forsberg M, L\u00a8onngren L. Saldo: a touch of Yin to wordnet\u2019s Yang. Lang Resour Evaluation. 2013;47:1191\u2013211.","journal-title":"Lang Resour Evaluation"},{"key":"4120_CR4","unstructured":"Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J. Distributed repre- sentations of words and phrases and their compositionality. Adv Neural Inf Process Syst 26 (2013)."},{"issue":"4","key":"4120_CR5","doi-asserted-by":"publisher","first-page":"397","DOI":"10.1007\/s42979-024-02688-6","volume":"5","author":"L Koopmans","year":"2024","unstructured":"Koopmans L, Dhali MA, Schomaker L. Performance analysis of handwritten text augmentation on style-based dating of historical documents. SN Comput Sci. 2024;5(4):397.","journal-title":"SN Comput Sci"},{"key":"4120_CR6","unstructured":"Zhang X, Zhao J, LeCun Y. Character-level convolutional networks for text classification. Adv Neural Inf Process Syst 28 (2015)."},{"key":"4120_CR7","unstructured":"Shleifer S. Low resource text classification with ulmfit and backtranslation. arXiv preprint arXiv:1903.09244 (2019)."},{"issue":"1","key":"4120_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s40649-020-00080-x","volume":"8","author":"H-T Duong","year":"2021","unstructured":"Duong H-T, Nguyen-Thi T-A. A review: preprocessing techniques and data augmentation for sentiment analysis. Comput Social Networks. 2021;8(1):1.","journal-title":"Comput Social Networks"},{"key":"4120_CR9","doi-asserted-by":"crossref","unstructured":"Wei J, Zou K. Eda: Easy data augmentation techniques for boosting perfor- mance on text classification tasks. arXiv preprint arXiv:1901.11196 (2019).","DOI":"10.18653\/v1\/D19-1670"},{"key":"4120_CR10","doi-asserted-by":"crossref","unstructured":"Kobayashi S. Contextual augmentation: data augmentation by words with paradigmatic relations. ArXiv Preprint arXiv:180506201 (2018).","DOI":"10.18653\/v1\/N18-2072"},{"key":"4120_CR11","doi-asserted-by":"crossref","unstructured":"Kumar V, Choudhary A, Cho E. Data augmentation using pre-trained transformer models. arXiv preprint arXiv:2003.02245 (2020).","DOI":"10.18653\/v1\/2020.lifelongnlp-1.3"},{"issue":"7","key":"4120_CR12","doi-asserted-by":"publisher","first-page":"2033","DOI":"10.1007\/s13042-021-01503-5","volume":"13","author":"AS Karnyoto","year":"2022","unstructured":"Karnyoto AS, Sun C, Liu B, Wang X. Augmentation and heterogeneous graph neural network for aaai2021-covid-19 fake news detection. Int J Mach Learn Cybernet. 2022;13(7):2033\u201343.","journal-title":"Int J Mach Learn Cybernet"},{"key":"4120_CR13","unstructured":"Schick T. Few-shot learning with language models: Learning from instructions and contexts. PhD thesis, lmu (2022)."},{"key":"4120_CR14","unstructured":"Zhang N, Li L, Chen X, Deng S, Bi Z, Tan C, Huang F, Chen H. Dif- ferentiable prompt makes pre-trained Language models better few-shot learners. ArXiv Preprint arXiv:210813161 (2021)."},{"key":"4120_CR15","doi-asserted-by":"publisher","unstructured":"Yang H, Li K. Boosting text augmentation via hybrid instance filtering frame- work. In: Findings of the Association for Computational Linguistics: ACL 2023, pp. 1652\u20131669. Association for Computational Linguistics, Toronto, Canada (2023). https:\/\/doi.org\/10.18653\/v1\/2023.findings-acl.105. https:\/\/aclanthology.org\/2023.findings-acl.105","DOI":"10.18653\/v1\/2023.findings-acl.105"},{"key":"4120_CR16","unstructured":"Authors M. Aeda: random punctuation-based augmentation for Nlp tasks. Com- Prehensive Study Text Augmentation Methods (2023). Detailed study available at ArXiv."},{"key":"4120_CR17","unstructured":"Team R. Iterative Mask Prediction: Contextual Augmentation with BERT (2023). https:\/\/arxiv.org\/abs\/2402.09141"},{"key":"4120_CR18","unstructured":"Team ANR. Conditional variational autoencoder for class-specific text aug- mentation (2023). Included in 2023 comprehensive studies."},{"key":"4120_CR19","unstructured":"Mahamud M, Lee Z, Samsten I. Distributional Data Augmentation Methods for Low Resource Language (2023). https:\/\/arxiv.org\/abs\/2309.04862"},{"key":"4120_CR20","unstructured":"Researchers V. Curriculum augmentation for improved text classification. In: NLP Advancements 2023 (2023). Findings published across recent NLP work- shops. https:\/\/arxiv.org\/abs\/2402.09141"},{"key":"4120_CR21","unstructured":"Schick T et al. Forward-looking active retrieval augmented generation (flare). arXiv preprint arXiv:2305.06983 (2023)."},{"key":"4120_CR22","unstructured":"Unknown A. Advancing nlp models with strategic text augmentation: A comprehensive study. arXiv preprint arXiv:2402.09141 (2023)."},{"key":"4120_CR23","unstructured":"Pingle A, Vyawahare A, Joshi I, Tangsali R, Kale G, Joshi R. Robust Sentiment Analysis for Low Resource languages Using Data Augmentation Approaches: A Case Study in Marathi (2023). https:\/\/arxiv.org\/abs\/2310.00734"},{"key":"4120_CR24","unstructured":"Litake O, Yagnik N, Labhsetwar S. IndiText Boost: Text Augmentation for Low Resource India Languages (2024). https:\/\/arxiv.org\/abs\/2401.13085"},{"key":"4120_CR25","doi-asserted-by":"crossref","unstructured":"Ehsan T, Solorio T. Enhancing NER performance in Low-Resource Pakistani languages using Cross-Lingual data augmentation (2025). https:\/\/arxiv.org\/abs\/2504.08792","DOI":"10.18653\/v1\/2025.wnut-1.13"},{"key":"4120_CR26","unstructured":"Adesam Y, Berdicevskis A, Morger F. Swedishglue\u2013towards a swedish test set for evaluating natural language understanding models (2020)."},{"key":"4120_CR27","unstructured":"Volodina E, Mohammed YA, Klezl J. Dalaj-a dataset for linguistic acceptability judgments for swedish: Format, baseline, sharing. arXiv preprint arXiv:2105.06681 (2021)."},{"issue":"1","key":"4120_CR28","doi-asserted-by":"publisher","first-page":"101","DOI":"10.1186\/s40537-021-00492-0","volume":"8","author":"C Shorten","year":"2021","unstructured":"Shorten C, Khoshgoftaar TM, Furht B. Text data augmentation for deep learning. J Big Data. 2021;8(1):101.","journal-title":"J Big Data"},{"issue":"1","key":"4120_CR29","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1162\/coli_a_00425","volume":"48","author":"GG S\u00b8ahin","year":"2022","unstructured":"S\u00b8ahin GG. To augment or not to augment? A comparative study on text aug- mentation techniques for low-resource Nlp. Comput Linguistics. 2022;48(1):5\u201342.","journal-title":"Comput Linguistics"},{"issue":"15","key":"4120_CR30","doi-asserted-by":"publisher","first-page":"17652","DOI":"10.1007\/s10489-022-03185-0","volume":"52","author":"C Tang","year":"2022","unstructured":"Tang C, Ma K, Cui B, Ji K, Abraham A. Long text feature extrac- Tion network with data augmentation. Appl Intell. 2022;52(15):17652\u201367.","journal-title":"Appl Intell"},{"issue":"4","key":"4120_CR31","doi-asserted-by":"publisher","first-page":"3129","DOI":"10.1007\/s12652-021-03439-8","volume":"14","author":"JJ Bird","year":"2023","unstructured":"Bird JJ, Ek\u00b4art A, Faria DR. Chatbot interaction with artificial intelligence: human data augmentation with t5 and Language transformer ensemble for text classification. J Ambient Intell Humaniz Comput. 2023;14(4):3129\u201344.","journal-title":"J Ambient Intell Humaniz Comput"},{"issue":"2","key":"4120_CR32","doi-asserted-by":"publisher","first-page":"399","DOI":"10.1007\/s10984-021-09368-9","volume":"25","author":"M Shardlow","year":"2022","unstructured":"Shardlow M, Sellar S, Rousell D. Collaborative augmentation and simplifi- cation of text (coast): pedagogical applications of natural Language processing in digital learning environments. Learn Environ Res. 2022;25(2):399\u2013421.","journal-title":"Learn Environ Res"}],"container-title":["SN Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-025-04120-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s42979-025-04120-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-025-04120-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,7]],"date-time":"2025-09-07T02:46:20Z","timestamp":1757213180000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s42979-025-04120-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,9]]},"references-count":32,"journal-issue":{"issue":"6","published-online":{"date-parts":[[2025,8]]}},"alternative-id":["4120"],"URL":"https:\/\/doi.org\/10.1007\/s42979-025-04120-z","relation":{},"ISSN":["2661-8907"],"issn-type":[{"type":"electronic","value":"2661-8907"}],"subject":[],"published":{"date-parts":[[2025,7,9]]},"assertion":[{"value":"3 June 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 June 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 July 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"There is no Conflict of Interest as of Our Knowledge.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}}],"article-number":"623"}}