{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T03:21:18Z","timestamp":1740108078589,"version":"3.37.3"},"reference-count":67,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,12,14]],"date-time":"2024-12-14T00:00:00Z","timestamp":1734134400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,14]],"date-time":"2024-12-14T00:00:00Z","timestamp":1734134400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"nrf"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2025,2]]},"DOI":"10.1007\/s00521-024-10663-8","type":"journal-article","created":{"date-parts":[[2024,12,14]],"date-time":"2024-12-14T08:24:17Z","timestamp":1734164657000},"page":"3473-3487","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Regulating the level of manipulation in text augmentation with systematic adjustment and advanced sentence embedding"],"prefix":"10.1007","volume":"37","author":[{"given":"Yuho","family":"Cha","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4199-936X","authenticated-orcid":false,"given":"Younghoon","family":"Lee","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,14]]},"reference":[{"issue":"1","key":"10663_CR1","doi-asserted-by":"publisher","first-page":"135","DOI":"10.1007\/s13042-022-01553-3","volume":"14","author":"M Bayer","year":"2023","unstructured":"Bayer M, Kaufhold M-A, Buchhold B, Keller M, Dallmeyer J, Reuter C (2023) Data augmentation in natural language processing: a novel text generation approach for long and short text classifiers. Int J Mach Learn Cybern 14(1):135\u2013150","journal-title":"Int J Mach Learn Cybern"},{"issue":"1","key":"10663_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3533048","volume":"17","author":"S Kwon","year":"2023","unstructured":"Kwon S, Lee Y (2023) Explainability-based mix-up approach for text data augmentation. ACM Trans Knowl Discov Data 17(1):1\u201314","journal-title":"ACM Trans Knowl Discov Data"},{"key":"10663_CR3","doi-asserted-by":"crossref","unstructured":"Feng SY, Gangal V, Wei J, Chandar S, Vosoughi S, Mitamura T, Hovy E (2021). A survey of data augmentation approaches for nlp. arXiv preprint arXiv:2105.03075","DOI":"10.18653\/v1\/2021.findings-acl.84"},{"key":"10663_CR4","doi-asserted-by":"crossref","unstructured":"Wei J, Zou K (2019) Eda: Easy data augmentation techniques for boosting performance on text classification tasks. arXiv preprint arXiv:1901.11196","DOI":"10.18653\/v1\/D19-1670"},{"key":"10663_CR5","doi-asserted-by":"crossref","unstructured":"Karimi A, Rossi L, Prati A (2021). Aeda: an easier data augmentation technique for text classification. arXiv preprint arXiv:2108.13230","DOI":"10.18653\/v1\/2021.findings-emnlp.234"},{"key":"10663_CR6","first-page":"6256","volume":"33","author":"Q Xie","year":"2020","unstructured":"Xie Q, Dai Z, Hovy E, Luong T, Le Q (2020) Unsupervised data augmentation for consistency training. Adv Neural Inf Process Syst 33:6256\u20136268","journal-title":"Adv Neural Inf Process Syst"},{"key":"10663_CR7","doi-asserted-by":"publisher","unstructured":"Chen H. Ji Y, Evans D (2020). Finding Friends and flipping frenemies: Automatic paraphrase dataset augmentation using graph theory. In: Cohn, T., He, Y., Liu, Y. (Eds.) Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 4741\u20134751. Association for Computational Linguistics, Online . https:\/\/doi.org\/10.18653\/v1\/2020.findings-emnlp.426 . https:\/\/aclanthology.org\/2020.findings-emnlp.426","DOI":"10.18653\/v1\/2020.findings-emnlp.426"},{"key":"10663_CR8","doi-asserted-by":"publisher","unstructured":"Guo D, Kim Y, Rush A (2020). Sequence-level mixed sample data augmentation. In: Webber B, Cohn T, He Y, Liu Y (ed) Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Association for Computational Linguistics, Online. pp. 5547\u20135552. https:\/\/doi.org\/10.18653\/v1\/2020.emnlp-main.447 . https:\/\/aclanthology.org\/2020.emnlp-main.447","DOI":"10.18653\/v1\/2020.emnlp-main.447"},{"key":"10663_CR9","doi-asserted-by":"publisher","unstructured":"Cai H, Chen H, Song Y, Zhang C, Zhao X, Yin D (2020) Data manipulation: Towards effective instance learning for neural dialogue generation via learning to augment and reweight. In: Jurafsky D, Chai J, Schluter N, Tetreault J (ed.) Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics, Online. pp 6334\u20136343. https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.564 . https:\/\/aclanthology.org\/2020.acl-main.564","DOI":"10.18653\/v1\/2020.acl-main.564"},{"key":"10663_CR10","doi-asserted-by":"publisher","unstructured":"Cheng Y, Jiang L, Macherey W, Eisenstein J (2020) AdvAug: Robust adversarial augmentation for neural machine translation. In: Jurafsky D, Chai, J., Schluter, N., Tetreault, J. (eds.) Proceedings of the 58th Annual Meeting of the Association for Computational LinguisticsAssociation for Computational Linguistics, Online . pp 5961\u20135970. https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.529 . https:\/\/aclanthology.org\/2020.acl-main.529","DOI":"10.18653\/v1\/2020.acl-main.529"},{"key":"10663_CR11","doi-asserted-by":"crossref","unstructured":"Gangal V, Feng SY, Alikhani M, Mitamura T, Hovy E (2022) Nareor: The narrative reordering problem. In: Proceedings of the AAAI Conference on Artificial Intelligence, 36:10645\u201310653","DOI":"10.1609\/aaai.v36i10.21309"},{"key":"10663_CR12","doi-asserted-by":"publisher","unstructured":"Thakur N, Reimers N, Daxenberger J, Gurevych I (2021). Augmented SBERT: Data augmentation method for improving bi-encoders for pairwise sentence scoring tasks. In: Toutanova K, Rumshisky A, Zettlemoyer L, Hakkani-Tur D, Beltagy I, Bethard S, Cotterell R, Chakraborty T, Zhou Y (ed) Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics, pp. 296\u2013310. Online. https:\/\/doi.org\/10.18653\/v1\/2021.naacl-main.28 . https:\/\/aclanthology.org\/2021.naacl-main.28","DOI":"10.18653\/v1\/2021.naacl-main.28"},{"key":"10663_CR13","doi-asserted-by":"publisher","unstructured":"Jindal A, Ranganatha NE, Didolkar A, Chowdhury AG, Jin D, Sawhney R, Shah RR (2020) SpeechMix-augmenting deep sound recognition using hidden space interpolations. In: Proc. Interspeech 2020, pp 861\u2013865 . https:\/\/doi.org\/10.21437\/Interspeech.2020-3147","DOI":"10.21437\/Interspeech.2020-3147"},{"key":"10663_CR14","doi-asserted-by":"crossref","unstructured":"Liu R, Xu G, Jia C, Ma W, Wang L, Vosoughi S (2020) Data boost: text data augmentation through reinforcement learning guided conditional generation. arXiv preprint arXiv:2012.02952","DOI":"10.18653\/v1\/2020.emnlp-main.726"},{"key":"10663_CR15","doi-asserted-by":"publisher","unstructured":"Yang W, Zhang R, Chen J, Wang L, Kim J (2023) Prototype-guided pseudo labeling for semi-supervised text classification. In: Rogers A, Boyd-Graber J, Okazaki N (ed) Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Association for Computational Linguistics, Toronto, Canada . pp. 16369\u201316382. https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.904 . https:\/\/aclanthology.org\/2023.acl-long.904","DOI":"10.18653\/v1\/2023.acl-long.904"},{"key":"10663_CR16","doi-asserted-by":"publisher","unstructured":"Liu R, Xu G, Jia C, Ma W, Wang L, Vosoughi S (2020) Data boost: text data augmentation through reinforcement learning guided conditional generation. In: Webber B, Cohn T, He Y, Liu Y. (ed) Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP),Association for Computational Linguistics, pp. 9031\u20139041. Online . https:\/\/doi.org\/10.18653\/v1\/2020.emnlp-main.726 . https:\/\/aclanthology.org\/2020.emnlp-main.726","DOI":"10.18653\/v1\/2020.emnlp-main.726"},{"key":"10663_CR17","doi-asserted-by":"publisher","unstructured":"Yang Y, Malaviya C, Fernandez J, Swayamdipta S, Le\u00a0Bras R, Wang JP, Bhagavatula C, Choi Y, Downey D (2020) Generative data augmentation for commonsense reasoning. In: Cohn T, He, Y, Liu Y (ed) Findings of the association for computational linguistics: EMNLP 2020, Association for Computational Linguistics, pp. 1008\u20131025. Online . https:\/\/doi.org\/10.18653\/v1\/2020.findings-emnlp.90 . https:\/\/aclanthology.org\/2020.findings-emnlp.90","DOI":"10.18653\/v1\/2020.findings-emnlp.90"},{"key":"10663_CR18","unstructured":"Hu Z, Tan B, Salakhutdinov R, Mitchell T, Xing EP (2019) Learning data manipulation for augmentation and weighting. Curran Associates Inc., Red Hook, NY, USA"},{"key":"10663_CR19","doi-asserted-by":"publisher","unstructured":"Li S, Yang M, Li C, Xu R (2022). Dual pseudo supervision for semi-supervised text classification with a reliable teacher. In: Proceedings of the 45th International ACM SIGIR conference on research and development in information retrieval. SIGIR \u201922, Association for Computing Machinery, New York, NY, USA . pp. 2513\u20132518.https:\/\/doi.org\/10.1145\/3477495.3531887","DOI":"10.1145\/3477495.3531887"},{"key":"10663_CR20","unstructured":"Devlin J, Chang MW, Lee K, Toutanova K (2019) Bert: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, Vol. 1 (Long and Short Papers), pp. 4171\u20134186"},{"key":"10663_CR21","doi-asserted-by":"crossref","unstructured":"Cornegruta S, Bakewell R, Withey S, Montana G (2016) Modelling radiological language with bidirectional long short-term memory networks. arXiv preprint arXiv:1609.08409","DOI":"10.18653\/v1\/W16-6103"},{"key":"10663_CR22","doi-asserted-by":"crossref","unstructured":"Xu B, Qiu S, Zhang J, Wang Y, Shen X, Melo G (2020) Data augmentation for multiclass utterance classification\u2013a systematic study. In: Proceedings of the 28th international conference on computational linguistics, pp 5494\u20135506","DOI":"10.18653\/v1\/2020.coling-main.479"},{"key":"10663_CR23","doi-asserted-by":"publisher","first-page":"105918","DOI":"10.1016\/j.knosys.2020.105918","volume":"197","author":"S Liu","year":"2020","unstructured":"Liu S, Lee K, Lee I (2020) Document-level multi-topic sentiment classification of email data with bilstm and data augmentation. Knowl Based Syst 197:105918","journal-title":"Knowl Based Syst"},{"key":"10663_CR24","doi-asserted-by":"crossref","unstructured":"Dai X, Adel H (2020) An analysis of simple data augmentation for named entity recognition. arXiv preprint arXiv:2010.11683","DOI":"10.18653\/v1\/2020.coling-main.343"},{"key":"10663_CR25","doi-asserted-by":"crossref","unstructured":"Marivate V, Sefara T (2020). Improving short text classification through global augmentation methods. In: Machine Learning and Knowledge Extraction: 4th IFIP TC 5, TC 12, WG 8.4, WG 8.9, WG 12.9 International Cross-Domain Conference, CD-MAKE 2020, Dublin, Ireland, August 25\u201328, 2020, Proceedings 4, Springer, pp 385\u2013399","DOI":"10.1007\/978-3-030-57321-8_21"},{"key":"10663_CR26","doi-asserted-by":"crossref","unstructured":"Gao F, Zhu J, Wu L, Xia Y, Qin T, Cheng X, Zhou W, Liu TY (2019) Soft contextual data augmentation for neural machine translation. In: Proceedings of the 57th annual meeting of the association for computational linguistics, Association for Computational Linguistics, Florence, pp 5539\u20135544","DOI":"10.18653\/v1\/P19-1555"},{"key":"10663_CR27","doi-asserted-by":"crossref","unstructured":"Wang X, Pham H, Dai Z, Neubig G (2018) Switchout: an efficient data augmentation algorithm for neural machine translation. arXiv preprint arXiv:1808.07512","DOI":"10.18653\/v1\/D18-1100"},{"key":"10663_CR28","doi-asserted-by":"crossref","unstructured":"Zhang R, Yu Y, Zhang C (2020) Seqmix: Augmenting active sequence labeling via sequence mixup. arXiv preprint arXiv:2010.02322","DOI":"10.18653\/v1\/2020.emnlp-main.691"},{"key":"10663_CR29","doi-asserted-by":"crossref","unstructured":"Ng N, Cho K, Ghassemi M (2020) Ssmba: Self-supervised manifold based data augmentation for improving out-of-domain robustness. arXiv preprint arXiv:2009.10195","DOI":"10.18653\/v1\/2020.emnlp-main.97"},{"key":"10663_CR30","doi-asserted-by":"crossref","unstructured":"Chen J, Yang Z, Yang D (2020) Mixtext: Linguistically-informed interpolation of hidden space for semi-supervised text classification. arXiv preprint arXiv:2004.12239","DOI":"10.18653\/v1\/2020.acl-main.194"},{"key":"10663_CR31","doi-asserted-by":"publisher","first-page":"8363","DOI":"10.1109\/ACCESS.2022.3142843","volume":"10","author":"M Kim","year":"2022","unstructured":"Kim M, Kang P (2022) Text embedding augmentation based on retraining with pseudo-labeled adversarial embedding. IEEE Access 10:8363\u20138376","journal-title":"IEEE Access"},{"key":"10663_CR32","doi-asserted-by":"crossref","unstructured":"Edunov S, Ott M, Auli M, Grangier D (2018) Understanding back-translation at scale. arXiv preprint arXiv:1808.09381","DOI":"10.18653\/v1\/D18-1045"},{"key":"10663_CR33","doi-asserted-by":"crossref","unstructured":"Hayashi T, Watanabe S, Zhang Y, Toda T, Hori T, Astudillo R, Takeda K (2018) Back-translation-style data augmentation for end-to-end asr. In: 2018 IEEE spoken language technology workshop (SLT), IEEE, pp 426\u2013433","DOI":"10.1109\/SLT.2018.8639619"},{"key":"10663_CR34","doi-asserted-by":"crossref","unstructured":"Ibrahim M, Torki M, El-Makky NM (2020) Alexu-backtranslation-tl at semeval-2020 task 12: Improving offensive language detection using data augmentation and transfer learning. In: Proceedings of the Fourteenth Workshop on Semantic Evaluation, pp 1881\u20131890","DOI":"10.18653\/v1\/2020.semeval-1.248"},{"key":"10663_CR35","doi-asserted-by":"crossref","unstructured":"Ding L, Wu D, Tao D (2021) Improving neural machine translation by bidirectional training. arXiv preprint arXiv:2109.07780","DOI":"10.18653\/v1\/2021.emnlp-main.263"},{"key":"10663_CR36","doi-asserted-by":"crossref","unstructured":"Sugiyama A, Yoshinaga N (2019) Data augmentation using back-translation for context-aware neural machine translation. In: Proceedings of the fourth workshop on discourse in machine translation (DiscoMT 2019), pp 35\u201344","DOI":"10.18653\/v1\/D19-6504"},{"key":"10663_CR37","doi-asserted-by":"crossref","unstructured":"Wu X, Lv S, Zang L, Han J, Hu S (2019) Conditional bert contextual augmentation. In: Computational science\u2013ICCS 2019: 19th international conference, Faro, Portugal, June 12\u201314, 2019, Proceedings, Part IV 19, 84\u201395 . Springer","DOI":"10.1007\/978-3-030-22747-0_7"},{"issue":"17","key":"10663_CR38","doi-asserted-by":"publisher","first-page":"5978","DOI":"10.3390\/app10175978","volume":"10","author":"V Atliha","year":"2020","unstructured":"Atliha V, \u0160e\u0161ok D (2020) Text augmentation using bert for image captioning. Appl Sci 10(17):5978","journal-title":"Appl Sci"},{"key":"10663_CR39","doi-asserted-by":"crossref","unstructured":"Yoo KM, Lee H, Dernoncourt F, Bui T, Chang W, Lee SG (2020) Variational hierarchical dialog autoencoder for dialog state tracking data augmentation. arXiv preprint arXiv:2001.08604","DOI":"10.18653\/v1\/2020.emnlp-main.274"},{"key":"10663_CR40","doi-asserted-by":"crossref","unstructured":"Anaby-Tavor A, Carmeli B, Goldbraich E, Kantor A, Kour G, Shlomov S, Tepper N, Zwerdling N (2020). Do not have enough data? deep learning to the rescue! In: Proceedings of the AAAI conference on artificial intelligence, vol. 34, pp 7383\u20137390","DOI":"10.1609\/aaai.v34i05.6233"},{"issue":"8","key":"10663_CR41","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford A, Wu J (2019) Rewon child, david luan, dario amodei, and ilya sutskever. Language models are unsupervised multitask learners. OpenAI blog 1(8):9","journal-title":"OpenAI blog"},{"key":"10663_CR42","unstructured":"Zhang J, Zhao Y, Saleh M, Liu P (2020) Pegasus: Pre-training with extracted gap-sentences for abstractive summarization. In: International conference on machine learning, PMLR, pp 11328\u201311339"},{"key":"10663_CR43","doi-asserted-by":"crossref","unstructured":"Kumar V, Choudhary A, Cho E (2020). Data augmentation using pre-trained transformer models. arXiv preprint arXiv:2003.02245","DOI":"10.18653\/v1\/2020.lifelongnlp-1.3"},{"issue":"2","key":"10663_CR44","doi-asserted-by":"publisher","first-page":"1741","DOI":"10.1007\/s11063-022-10961-z","volume":"55","author":"J Yu","year":"2023","unstructured":"Yu J, Choi J, Lee Y (2023) Mixing approach for text data augmentation based on an ensemble of explainable artificial intelligence methods. Neural Process Lett 55(2):1741\u20131757","journal-title":"Neural Process Lett"},{"key":"10663_CR45","unstructured":"Lee DH et al (2013) Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In: Workshop on Challenges in Representation Learning, ICML, Atlanta, vol. 3, p. 896 ."},{"key":"10663_CR46","unstructured":"Berthelot D, Carlini N, Goodfellow I, Papernot N, Oliver A, Raffel CA (2019) Mixmatch: A holistic approach to semi-supervised learning. Advances in neural information processing systems, vol. 32"},{"key":"10663_CR47","doi-asserted-by":"crossref","unstructured":"Shim H, Luca S, Lowet D, Vanrumste B (2020) Data augmentation and semi-supervised learning for deep neural networks-based text classifier. In: Proceedings of the 35th annual ACM symposium on applied computing, pp 1119\u20131126","DOI":"10.1145\/3341105.3373992"},{"issue":"4","key":"10663_CR48","doi-asserted-by":"publisher","first-page":"549","DOI":"10.1016\/j.icte.2021.11.002","volume":"8","author":"YJ Yu","year":"2022","unstructured":"Yu YJ, Yoon SJ, Jun SY, Kim JW (2022) Tabas: Text augmentation based on attention score for text classification model. ICT Express 8(4):549\u2013554","journal-title":"ICT Express"},{"key":"10663_CR49","doi-asserted-by":"crossref","unstructured":"Wagner J, Kohler JM, Gindele T, Hetzel L, Wiedemer JT, Behnke S (2019) Interpretable and fine-grained visual explanations for convolutional neural networks. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR)","DOI":"10.1109\/CVPR.2019.00931"},{"key":"10663_CR50","doi-asserted-by":"crossref","unstructured":"Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A (2016) Learning deep features for discriminative localization. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2921\u20132929","DOI":"10.1109\/CVPR.2016.319"},{"key":"10663_CR51","doi-asserted-by":"crossref","unstructured":"Chattopadhay A, Sarkar A, Howlader P, Balasubramanian VN (2018) Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks. In: 2018 IEEE winter conference on applications of computer vision (WACV), IEEE, pp 839\u2013847","DOI":"10.1109\/WACV.2018.00097"},{"key":"10663_CR52","unstructured":"Montavon G, Binder A, Lapuschkin S, Samek W, M\u00fcller K (2019) Explainable ai: interpreting, explaining and visualizing deep learning. Spring er LNCS, Vol. 11700"},{"key":"10663_CR53","unstructured":"Smilkov D, Thorat N, Kim B, Vi\u00e9gas F, Wattenberg M (2017) Smoothgrad: removing noise by adding noise. arXiv preprint arXiv:1706.03825"},{"key":"10663_CR54","doi-asserted-by":"crossref","unstructured":"Ribeiro MT, Singh S, Guestrin C (2016) why should i trust you? explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pp 1135\u20131144","DOI":"10.1145\/2939672.2939778"},{"key":"10663_CR55","unstructured":"Lundberg SM, Lee SI (2017). A unified approach to interpreting model predictions. Advances in neural information processing systems. vol. 30"},{"key":"10663_CR56","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiolchem.2022.107732","volume":"99","author":"NQK Le","year":"2022","unstructured":"Le NQK, Ho Q-T, Nguyen V-N, Chang J-S (2022) Bert-promoter: an improved sequence-based predictor of dna promoter using bert pre-trained model and shap feature selection. Comput Biol Chem 99:107732","journal-title":"Comput Biol Chem"},{"issue":"1","key":"10663_CR57","doi-asserted-by":"publisher","first-page":"802","DOI":"10.1109\/TVCG.2021.3114794","volume":"28","author":"X Wang","year":"2021","unstructured":"Wang X, He J, Jin Z, Yang M, Wang Y, Qu H (2021) M2lens: Visualizing and explaining multimodal models for sentiment analysis. IEEE Trans Visual Comput Gr 28(1):802\u2013812","journal-title":"IEEE Trans Visual Comput Gr"},{"issue":"11","key":"10663_CR58","doi-asserted-by":"publisher","first-page":"2673","DOI":"10.1109\/78.650093","volume":"45","author":"M Schuster","year":"1997","unstructured":"Schuster M, Paliwal KK (1997) Bidirectional recurrent neural networks. IEEE Trans Signal Process 45(11):2673\u20132681. https:\/\/doi.org\/10.1109\/78.650093","journal-title":"IEEE Trans Signal Process"},{"key":"10663_CR59","doi-asserted-by":"publisher","unstructured":"Lee JY, Dernoncourt F (2016) Sequential short-text classification with recurrent and convolutional neural networks. In: Knight K, Nenkova A, Rambow O (ed.) Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: Human language technologies, Association for Computational Linguistics, San Diego, California. pp 515\u2013520. https:\/\/doi.org\/10.18653\/v1\/N16-1062 . https:\/\/aclanthology.org\/N16-1062","DOI":"10.18653\/v1\/N16-1062"},{"issue":"4","key":"10663_CR60","doi-asserted-by":"publisher","first-page":"369","DOI":"10.1504\/IJCAT.2022.125186","volume":"68","author":"L Enamoto","year":"2022","unstructured":"Enamoto L, Santos AR, Maia R, Weigang L, Filho GPR (2022) Multi-label legal text classification with bilstm and attention. Int J Comput Appl Technol 68(4):369\u2013378","journal-title":"Int J Comput Appl Technol"},{"key":"10663_CR61","doi-asserted-by":"crossref","unstructured":"Duan A, Raga RC (2024) Bilstm model with attention mechanism for multi-label news text classification. In: 2024 4th International conference on neural networks, information and communication (NNICE), IEEE, pp 566\u2013569","DOI":"10.1109\/NNICE61279.2024.10498894"},{"key":"10663_CR62","unstructured":"Bird S, Klein E, Loper E (2009). Natural Language Processing with Python: Analyzing Text with the Natural Language Toolkit. O\u2019Reilly Media, Inc., ???"},{"key":"10663_CR63","unstructured":"Maas A, Daly RE, Pham PT, Huang D, Ng AY, Potts C (2011) Learning word vectors for sentiment analysis. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp 142\u2013150"},{"key":"10663_CR64","doi-asserted-by":"crossref","unstructured":"Rosenthal S, Farra N, Nakov P (2019) Semeval-2017 task 4: Sentiment analysis in twitter. arXiv preprint arXiv:1912.00741","DOI":"10.18653\/v1\/S17-2088"},{"key":"10663_CR65","unstructured":"Zhang X, Zhao J, LeCun Y (2015). Character-level convolutional networks for text classification. Advances in neural information processing systems 28"},{"key":"10663_CR66","doi-asserted-by":"publisher","unstructured":"Rafay A, Suleman M, Alim A (2020) Robust review rating prediction model based on machine and deep learning: Yelp dataset. In: 2020 International conference on emerging trends in smart technologies (ICETST), pp 8138\u20138143 . https:\/\/doi.org\/10.1109\/ICETST49965.2020.9080713","DOI":"10.1109\/ICETST49965.2020.9080713"},{"key":"10663_CR67","unstructured":"Twitter US Airline Sentiment. (2015) https:\/\/www.kaggle.com\/crowdflower\/twitter-airline-sentiment"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-024-10663-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-024-10663-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-024-10663-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,7]],"date-time":"2025-02-07T09:03:10Z","timestamp":1738918990000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-024-10663-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,14]]},"references-count":67,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["10663"],"URL":"https:\/\/doi.org\/10.1007\/s00521-024-10663-8","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"type":"print","value":"0941-0643"},{"type":"electronic","value":"1433-3058"}],"subject":[],"published":{"date-parts":[[2024,12,14]]},"assertion":[{"value":"29 January 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 October 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 December 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that the authors have no conflict of interest as defined by Springer.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}