{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T17:19:50Z","timestamp":1771003190002,"version":"3.50.1"},"reference-count":30,"publisher":"SAGE Publications","issue":"1","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/journals.sagepub.com\/page\/policies\/text-and-data-mining-license"}],"content-domain":{"domain":["journals.sagepub.com"],"crossmark-restriction":true},"short-container-title":["Journal of Computational Methods in Sciences and Engineering"],"published-print":{"date-parts":[[2025,1]]},"abstract":"<jats:p>The rapid development of big data and artificial intelligence has made text topic classification an important part of natural language processing research, and it has also promoted the optimization of pre-trained model performance. In order to better promote the application of pre-trained models and improve the effect of text topic classification, this paper introduces the BERT (Bidirectional Encoder Representations from Transformer) model to conduct an in-depth exploration of English text topic classification. The text preprocesses the English text dataset through operations such as denoising, converting to lowercase, and removing stops, and then uses synonymous substitution to enhance the English text data. Subsequently, the BERT model was pre-trained, and the model was optimized and a BERT-based model structure was designed, followed by the construction of a topic classifier. Finally, this article also evaluated the practical effectiveness of the BERT-based model in English text topic classification. The research results show that when the classification number is 5, the BERT-based model can achieve the highest accuracy of 96.49%; when the number of tests is 50, the recall rate and F1 value of the BERT-based model are 96.10% and 91.66%, respectively, when the classification number is 5. The research results indicate that applying the BERT-based model to English text topic classification is completely feasible. It can improve its accuracy and recall, reduce classification time, and improve classification performance. Applying it to text classification can better improve the efficiency of text classification.<\/jats:p>","DOI":"10.1177\/14727978251321982","type":"journal-article","created":{"date-parts":[[2025,3,4]],"date-time":"2025-03-04T18:35:56Z","timestamp":1741113356000},"page":"669-684","update-policy":"https:\/\/doi.org\/10.1177\/sage-journals-update-policy","source":"Crossref","is-referenced-by-count":1,"title":["English text topic classification using BERT-based model"],"prefix":"10.1177","volume":"25","author":[{"given":"Xi","family":"Li","sequence":"first","affiliation":[{"name":"Shaoyang University"}]},{"given":"Lili","family":"Jia","sequence":"additional","affiliation":[{"name":"Huanghuai University"}]}],"member":"179","published-online":{"date-parts":[[2025,3,4]]},"reference":[{"key":"e_1_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.1145\/3495162"},{"key":"e_1_3_2_3_2","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2020.2992485"},{"key":"e_1_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.30595\/juita.v10i2.13262"},{"key":"e_1_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.30865\/mib.v4i4.2276"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10660-022-09560-w"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.11925\/infotech.2096-3467.2020.0965"},{"key":"e_1_3_2_8_2","doi-asserted-by":"publisher","DOI":"10.1145\/3439726"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1017\/pan.2021.37"},{"key":"e_1_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-018-09677-1"},{"key":"e_1_3_2_11_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijresmar.2018.09.009"},{"key":"e_1_3_2_12_2","doi-asserted-by":"publisher","DOI":"10.1145\/3530260"},{"key":"e_1_3_2_13_2","doi-asserted-by":"publisher","DOI":"10.3969\/j.issn.1672-9722.2020.05.021"},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.3390\/app12115720"},{"key":"e_1_3_2_15_2","doi-asserted-by":"publisher","DOI":"10.47852\/bonviewJCCE3202838"},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.3969\/j.issn.2097-1915.2023.01.015"},{"issue":"4","key":"e_1_3_2_17_2","first-page":"446","article-title":"Text classification of defects in power grid equipment based on the BERT pre-trained language model","volume":"44","author":"Yuan T","year":"2020","unstructured":"Yuan T, Ye Y, Liu H, et al. Text classification of defects in power grid equipment based on the BERT pre-trained language model. J Nanjing Univ Sci Technol (Nat Sci) 2020; 44(4): 446\u2013453.","journal-title":"J Nanjing Univ Sci Technol (Nat Sci)"},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.1177\/1094428120971683"},{"key":"e_1_3_2_19_2","doi-asserted-by":"publisher","DOI":"10.1017\/S1351324922000213"},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.1007\/s42417-022-00844-x"},{"key":"e_1_3_2_21_2","doi-asserted-by":"publisher","DOI":"10.3901\/JME.2021.23.106"},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.11925\/infotech.2096-3467.2022.0919"},{"issue":"9","key":"e_1_3_2_23_2","first-page":"46","article-title":"A data enhancement strategy for automatic summary tasks for long text and small data sets","volume":"36","author":"Zhou P","year":"2022","unstructured":"Zhou P, Xuefeng X, Cui Z, et al. A data enhancement strategy for automatic summary tasks for long text and small data sets. Chinese Journal of Informatics 2022; 36(9): 46\u201356.","journal-title":"Chinese Journal of Informatics"},{"key":"e_1_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.1007\/s13369-021-05780-8"},{"key":"e_1_3_2_25_2","doi-asserted-by":"publisher","DOI":"10.3390\/info13020067"},{"key":"e_1_3_2_26_2","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btab712"},{"key":"e_1_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.14801\/jaitc.2020.10.1.45"},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-021-09958-2"},{"key":"e_1_3_2_29_2","doi-asserted-by":"publisher","DOI":"10.1287\/ijds.2022.0016"},{"issue":"6","key":"e_1_3_2_30_2","first-page":"65","article-title":"Domain entity recognition based on pre-trained BERT word embedding model","volume":"5","author":"Ding L","year":"2019","unstructured":"Ding L, Wen W, Lin Q. Domain entity recognition based on pre-trained BERT word embedding model. Intelligence Engineering 2019; 5(6): 65\u201374.","journal-title":"Intelligence Engineering"},{"key":"e_1_3_2_31_2","doi-asserted-by":"publisher","DOI":"10.3969\/j.issn.1671-0673.2021.03.013"}],"container-title":["Journal of Computational Methods in Sciences and Engineering"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.1177\/14727978251321982","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/full-xml\/10.1177\/14727978251321982","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.1177\/14727978251321982","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T16:31:38Z","timestamp":1771000298000},"score":1,"resource":{"primary":{"URL":"https:\/\/journals.sagepub.com\/doi\/10.1177\/14727978251321982"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1]]},"references-count":30,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,1]]}},"alternative-id":["10.1177\/14727978251321982"],"URL":"https:\/\/doi.org\/10.1177\/14727978251321982","relation":{},"ISSN":["1472-7978","1875-8983"],"issn-type":[{"value":"1472-7978","type":"print"},{"value":"1875-8983","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1]]}}}