{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T17:51:25Z","timestamp":1770745885755,"version":"3.49.0"},"reference-count":23,"publisher":"Walter de Gruyter GmbH","issue":"4","license":[{"start":{"date-parts":[[2019,11,1]],"date-time":"2019-11-01T00:00:00Z","timestamp":1572566400000},"content-version":"unspecified","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,11,1]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:sec id=\"j_jdis-2019-0020_s_006\">\n                    <jats:title>Purpose<\/jats:title>\n                    <jats:p>Move recognition in scientific abstracts is an NLP task of classifying sentences of the abstracts into different types of language units. To improve the performance of move recognition in scientific abstracts, a novel model of move recognition is proposed that outperforms the BERT-based method.<\/jats:p>\n                  <\/jats:sec>\n                  <jats:sec id=\"j_jdis-2019-0020_s_007\">\n                    <jats:title>Design\/methodology\/approach<\/jats:title>\n                    <jats:p>Prevalent models based on BERT for sentence classification often classify sentences without considering the context of the sentences. In this paper, inspired by the BERT masked language model (MLM), we propose a novel model called the masked sentence model that integrates the content and contextual information of the sentences in move recognition. Experiments are conducted on the benchmark dataset PubMed 20K RCT in three steps. Then, we compare our model with HSLN-RNN, BERT-based and SciBERT using the same dataset.<\/jats:p>\n                  <\/jats:sec>\n                  <jats:sec id=\"j_jdis-2019-0020_s_008\">\n                    <jats:title>Findings<\/jats:title>\n                    <jats:p>Compared with the BERT-based and SciBERT models, the F1 score of our model outperforms them by 4.96% and 4.34%, respectively, which shows the feasibility and effectiveness of the novel model and the result of our model comes closest to the state-of-the-art results of HSLN-RNN at present.<\/jats:p>\n                  <\/jats:sec>\n                  <jats:sec id=\"j_jdis-2019-0020_s_009\">\n                    <jats:title>Research limitations<\/jats:title>\n                    <jats:p>The sequential features of move labels are not considered, which might be one of the reasons why HSLN-RNN has better performance. Our model is restricted to dealing with biomedical English literature because we use a dataset from PubMed, which is a typical biomedical database, to fine-tune our model.<\/jats:p>\n                  <\/jats:sec>\n                  <jats:sec id=\"j_jdis-2019-0020_s_010\">\n                    <jats:title>Practical implications<\/jats:title>\n                    <jats:p>The proposed model is better and simpler in identifying move structures in scientific abstracts and is worthy of text classification experiments for capturing contextual features of sentences.<\/jats:p>\n                  <\/jats:sec>\n                  <jats:sec id=\"j_jdis-2019-0020_s_011\">\n                    <jats:title>Originality\/value<\/jats:title>\n                    <jats:p>T he study proposes a masked sentence model based on BERT that considers the contextual features of the sentences in abstracts in a new way. The performance of this classification model is significantly improved by rebuilding the input layer without changing the structure of neural networks.<\/jats:p>\n                  <\/jats:sec>","DOI":"10.2478\/jdis-2019-0020","type":"journal-article","created":{"date-parts":[[2019,12,29]],"date-time":"2019-12-29T04:31:26Z","timestamp":1577593886000},"page":"42-55","source":"Crossref","is-referenced-by-count":6,"title":["Masked Sentence Model Based on BERT for Move Recognition in Medical Scientific Abstracts"],"prefix":"10.2478","volume":"4","author":[{"given":"Gaihong","family":"Yu","sequence":"first","affiliation":[{"name":"National Science Library, Chinese Academy of Sciences , Beijing 100190 , China"},{"name":"University of Chinese Academy of Sciences , Beijing 100049 , China"}]},{"given":"Zhixiong","family":"Zhang","sequence":"additional","affiliation":[{"name":"National Science Library, Chinese Academy of Sciences , Beijing 100190 , China"},{"name":"University of Chinese Academy of Sciences , Beijing 100049 , China"},{"name":"Wuhan Library, Chinese Academy of Sciences , Wuhan 430071 , China"}]},{"given":"Huan","family":"Liu","sequence":"additional","affiliation":[{"name":"National Science Library, Chinese Academy of Sciences , Beijing 100190 , China"},{"name":"University of Chinese Academy of Sciences , Beijing 100049 , China"}]},{"given":"Liangping","family":"Ding","sequence":"additional","affiliation":[{"name":"National Science Library, Chinese Academy of Sciences , Beijing 100190 , China"},{"name":"University of Chinese Academy of Sciences , Beijing 100049 , China"}]}],"member":"374","published-online":{"date-parts":[[2019,12,27]]},"reference":[{"key":"2026012108095965739_j_jdis-2019-0020_ref_001","unstructured":"Amini, I., Martinez, D., & Molla, D. (2012). Overview of the ALTA 2012 shared task. In Proceedings of the Australasian Language Technology Association Workshop 2012: ALTA 2012 (pp. 124\u2013129). Dunedin, New Zealand."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_002","doi-asserted-by":"crossref","unstructured":"Badie, K., Asadi, N., & Tayefeh Mahmoudi, M. (2018). Zone identification based on features with high semantic richness and combining results of separate classifiers. Journal of Information and Telecommunication, 2(4), 411\u2013427.","DOI":"10.1080\/24751839.2018.1460083"},{"key":"2026012108095965739_j_jdis-2019-0020_ref_003","doi-asserted-by":"crossref","unstructured":"Basili, R. & Pennacchiotti, M. (2010). Distributional lexical semantics: Toward uniform representation paradigms for advanced acquisition and processing tasks. Natural Language Engineering, 1(1), 1\u201312.","DOI":"10.1017\/S1351324910000112"},{"key":"2026012108095965739_j_jdis-2019-0020_ref_004","unstructured":"Beltagy, I., Lo, K., & Cohan, A. (2019). SciBERT: Pretrained contextualized embeddings for scientific text. arXiv:1903.10676v3."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_005","unstructured":"Dasigi, P., Burns, G.A.P.C., Hovy, E., & Waard, A. (2017). Experiment segmentation in scientific discourse as clause-level structured prediction using recurrent neural networks. arXiv:1702.05398."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_006","unstructured":"Devlin, J., Chang, M.W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_007","unstructured":"Ding, L.P., Zhang, Z.X., & Liu, H. (2019). Research on factors affecting the SVM model performance on move recognition. Data Analysis and Konwledge Discovery, http:\/\/kns.cnki.net\/kcms\/detail\/10.1478.G2.20191012.0931.002.html."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_008","unstructured":"Firth, J.R. (1930). A synopsis of linguistic theory, 1930\u20131955. In: Firth, J.R., Ed., Studies in Linguistic Analysis, Longmans, London, 168\u2013205."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_009","unstructured":"Fisas, B., Ronzano, F., & Saggion, H. (2016). A multi-layered annotated corpus of scientific papers. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC 2016)."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_010","unstructured":"Franck Dernoncourt & Ji Young Lee. (2017). Pubmed 200k rct: a dataset for sequential sentence classification in medical abstracts. In Proceedings of the 8th International Joint Conference on Natural Language Processing."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_011","doi-asserted-by":"crossref","unstructured":"Gerlach, M., Peixoto, T.P., Altmann, E.G., & Altmann, E.G. (2018). A network approach to topic models. Science advances, 4(7), eaaq1360.","DOI":"10.1126\/sciadv.aaq1360"},{"key":"2026012108095965739_j_jdis-2019-0020_ref_012","unstructured":"Hirohata, K., Okazaki, N., Ananiadou, S., & Mitsuru. (2018). Identifying sections in scientific abstracts using conditional random fields. In Proceedings of the Third International Joint Conference on Natural Language Processing."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_013","doi-asserted-by":"crossref","unstructured":"Ma, M.B., Huang, L., Xiang, B., & Zhou, B.W. (2015). Dependency-based convolutional neural networks for sentence embedding. arXiv:1507.01839.","DOI":"10.3115\/v1\/P15-2029"},{"key":"2026012108095965739_j_jdis-2019-0020_ref_014","doi-asserted-by":"crossref","unstructured":"Peters, M.E., Neumann, M., Iyyer, M., et al. (2018). Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. doi: 10.18653\/v1\/N18-1202 arXiv:1802.05365.","DOI":"10.18653\/v1\/N18-1202"},{"key":"2026012108095965739_j_jdis-2019-0020_ref_015","unstructured":"Radford, A., Narasimhan, K., Salimans, T., & Sutskever Ilya (2018). Improving language understanding by generative pre-training. https:\/\/s3-us-west-2.amazonaws.com\/openai-assets\/researchcovers\/languageunsupervised\/languageunderstandingpaper.pdf"},{"key":"2026012108095965739_j_jdis-2019-0020_ref_016","doi-asserted-by":"crossref","unstructured":"Lai, S.W., Xu, L., Liu, K., & Zhao, J. (2015). Recurrent convolutional neural networks for text classification. In AAAI\u201915 Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, pages 2267\u20132273.","DOI":"10.1609\/aaai.v29i1.9513"},{"key":"2026012108095965739_j_jdis-2019-0020_ref_017","doi-asserted-by":"crossref","unstructured":"Swales, J.M. (2004). Research genres: Explorations and applications. Cambridge: Cambridge University Press.","DOI":"10.1017\/CBO9781139524827"},{"key":"2026012108095965739_j_jdis-2019-0020_ref_018","doi-asserted-by":"crossref","unstructured":"Taylor, W.L. (1953). \u201cCloze procedure\u201d: A new tool for measuring readability. Journalism & Mass Communication Quarterly, 30(4), 415\u2013433. doi: https:\/\/doi.org\/10.1177\/107769905303000401","DOI":"10.1177\/107769905303000401"},{"key":"2026012108095965739_j_jdis-2019-0020_ref_019","unstructured":"Teufel, S. (1999). Argumentative zoning: Information extraction from scientific text. Edinburgh: University of Edinburgh."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_020","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., et al. (2017). Attention is all you need. arXiv:1706.03762v5."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_021","unstructured":"Yamamoto, Y. & Takagi, T. (2005). A sentence classification system for multi-document summarization in the biomedical domain. In Proceedings of International Workshop on Biomedical Data Engineering, pages 90\u201395."},{"key":"2026012108095965739_j_jdis-2019-0020_ref_022","doi-asserted-by":"crossref","unstructured":"Yoon Kim. (2014). Convolutional neural networks for sentence classification. arXiv:1408.5882.","DOI":"10.3115\/v1\/D14-1181"},{"key":"2026012108095965739_j_jdis-2019-0020_ref_023","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Liu, H., Ding, L., et al. (2019). Moves recognition in abstract of research paper based on deep learning. In Proceedings of 2019 ACM\/IEEE Joint Conference on Digital Libraries (JCDL). IEEE, pages 390\u2013391.","DOI":"10.1109\/JCDL.2019.00085"}],"container-title":["Journal of Data and Information Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/content.sciendo.com\/view\/journals\/jdis\/4\/4\/article-p42.xml","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.sciendo.com\/pdf\/10.2478\/jdis-2019-0020","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T08:10:10Z","timestamp":1768983010000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.sciendo.com\/article\/10.2478\/jdis-2019-0020"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,11,1]]},"references-count":23,"journal-issue":{"issue":"4","published-online":{"date-parts":[[2019,12,27]]},"published-print":{"date-parts":[[2019,11,1]]}},"alternative-id":["10.2478\/jdis-2019-0020"],"URL":"https:\/\/doi.org\/10.2478\/jdis-2019-0020","relation":{},"ISSN":["2543-683X"],"issn-type":[{"value":"2543-683X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019,11,1]]}}}