{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T06:10:02Z","timestamp":1770271802658,"version":"3.49.0"},"reference-count":48,"publisher":"MDPI AG","issue":"8","license":[{"start":{"date-parts":[[2022,4,8]],"date-time":"2022-04-08T00:00:00Z","timestamp":1649376000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and development Program","doi-asserted-by":"publisher","award":["2018YFB1004502"],"award-info":[{"award-number":["2018YFB1004502"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Named entity recognition (NER) is a task that seeks to recognize entities in raw texts and is a precondition for a series of downstream NLP tasks. Traditionally, prior NER models use the sequence labeling mechanism which requires label dependency captured by the conditional random fields (CRFs). However, these models are prone to cascade label misclassifications since a misclassified label results in incorrect label dependency, and so some following labels may also be misclassified. To address the above issue, we propose S-NER, a span-based NER model. To be specific, S-NER first splits raw texts into text spans and regards them as candidate entities; it then directly obtains the types of spans by conducting entity type classifications on span semantic representations, which eliminates the requirement for label dependency. Moreover, S-NER has a concise neural architecture in which it directly uses BERT as its encoder and a feed-forward network as its decoder. We evaluate S-NER on several benchmark datasets across three domains. Experimental results demonstrate that S-NER consistently outperforms the strongest baselines in terms of F1-score. Extensive analyses further confirm the efficacy of S-NER.<\/jats:p>","DOI":"10.3390\/s22082852","type":"journal-article","created":{"date-parts":[[2022,4,9]],"date-time":"2022-04-09T05:13:08Z","timestamp":1649481188000},"page":"2852","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":20,"title":["S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition"],"prefix":"10.3390","volume":"22","author":[{"given":"Jie","family":"Yu","sequence":"first","affiliation":[{"name":"College of Computer, National University of Defense Technology, Changsha 410073, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5508-5051","authenticated-orcid":false,"given":"Bin","family":"Ji","sequence":"additional","affiliation":[{"name":"College of Computer, National University of Defense Technology, Changsha 410073, China"}]},{"given":"Shasha","family":"Li","sequence":"additional","affiliation":[{"name":"College of Computer, National University of Defense Technology, Changsha 410073, China"}]},{"given":"Jun","family":"Ma","sequence":"additional","affiliation":[{"name":"College of Computer, National University of Defense Technology, Changsha 410073, China"}]},{"given":"Huijun","family":"Liu","sequence":"additional","affiliation":[{"name":"College of Computer, National University of Defense Technology, Changsha 410073, China"}]},{"given":"Hao","family":"Xu","sequence":"additional","affiliation":[{"name":"College of Computer, National University of Defense Technology, Changsha 410073, China"}]}],"member":"1968","published-online":{"date-parts":[[2022,4,8]]},"reference":[{"key":"ref_1","first-page":"8401","article-title":"HAMNER: Headword amplified multi-span distantly supervised method for domain specific named entity recognition","volume":"34","author":"Liu","year":"2020","journal-title":"AAAI Conf. Artif. Intell."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Jie, Z., and Lu, W. (2019, January 3\u20137). Dependency-Guided LSTM-CRF for Named Entity Recognition. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China.","DOI":"10.18653\/v1\/D19-1399"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Sui, D., Chen, Y., Liu, K., Zhao, J., and Liu, S. (2019, January 3\u20137). Leverage lexical knowledge for Chinese named entity recognition via collaborative graph network. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China.","DOI":"10.18653\/v1\/D19-1396"},{"key":"ref_4","unstructured":"Yan, H., Deng, B., Li, X., and Qiu, X. (2019). TENER: Adapting Transformer Encoder for Named Entity Recognition. arXiv."},{"key":"ref_5","unstructured":"Huang, Z., Xu, W., and Yu, K. (2015). Bidirectional LSTM-CRF models for sequence tagging. arXiv."},{"key":"ref_6","unstructured":"Lafferty, J.D., McCallum, A., and Pereira, F.C. (July, January 28). Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data. Proceedings of theProceedings of the 18th International Conference on Machine Learning 2001 (ICML 2001), Williamstown, MA, USA."},{"key":"ref_7","first-page":"9016","article-title":"Boundary enhanced neural span classification for nested named entity recognition","volume":"34","author":"Tan","year":"2020","journal-title":"AAAI Conf. Artif. Intell."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"135091","DOI":"10.1109\/ACCESS.2020.3011598","article-title":"A bidirectional iterative algorithm for nested named entity recognition","volume":"8","author":"Dadas","year":"2020","journal-title":"IEEE Access"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Li, F., Lin, Z., Zhang, M., and Ji, D. (2021, January 1\u20136). A Span-Based Model for Joint Overlapped and Discontinuous Named Entity Recognition. Proceedings of the Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Online.","DOI":"10.18653\/v1\/2021.acl-long.372"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Luan, Y., Wadden, D., He, L., Shah, A., Ostendorf, M., and Hajishirzi, H. (2019, January 2\u20137). A general framework for information extraction using dynamic span graphs. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA.","DOI":"10.18653\/v1\/N19-1308"},{"key":"ref_11","unstructured":"Dixit, K., and Al-Onaizan, Y. (August, January 28). Span-level model for relation extraction. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Ouchi, H., Suzuki, J., Kobayashi, S., Yokoi, S., Kuribayashi, T., Konno, R., and Inui, K. (2020, January 5\u201310). Instance-Based Learning of Span Representations: A Case Study through Named Entity Recognition. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online.","DOI":"10.18653\/v1\/2020.acl-main.575"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Lin, B.Y., Xu, F.F., Luo, Z., and Zhu, K. (2017, January 7\u20139). Multi-channel bilstm-crf model for emerging named entity recognition in social media. Proceedings of the 3rd Workshop on Noisy User-generated Text, Copenhagen, Denmark.","DOI":"10.18653\/v1\/W17-4421"},{"key":"ref_14","unstructured":"Zheng, S., Wang, F., Bao, H., Hao, Y., Zhou, P., and Xu, B. (August, January 30). Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Chen, J., Yuan, C., Wang, X., and Bai, Z. (2019, January 3\u20134). MrMep: Joint extraction of multiple relations and multiple entity pairs based on triplet attention. Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), Hong Kong, China.","DOI":"10.18653\/v1\/K19-1055"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Zhou, P., Zheng, S., Xu, J., Qi, Z., Bao, H., and Xu, B. (2017). Joint extraction of multiple relations and entities by using a hybrid neural network. Chinese Computational Linguistics and Natural Language Processing Based on Naturally Annotated Big Data, Springer.","DOI":"10.1007\/978-3-319-69005-6_12"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Ye, H., Zhang, N., Deng, S., Chen, M., Tan, C., Huang, F., and Chen, H. (2020). Contrastive Triple Extraction with Generative Transformer. arXiv.","DOI":"10.1109\/TASLP.2021.3110126"},{"key":"ref_18","unstructured":"Kenton, J.D.M.W.C., and Toutanova, L.K. (2019, January 2\u20137). BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA."},{"key":"ref_19","unstructured":"Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., and Stoyanov, V. (2019). RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Fu, J., Huang, X., and Liu, P. (2021, January 1\u20136). SpanNER: Named Entity Re-\/Recognition as Span Prediction. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online.","DOI":"10.18653\/v1\/2021.acl-long.558"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","article-title":"Long short-term memory","volume":"9","author":"Hochreiter","year":"1997","journal-title":"Neural Comput."},{"key":"ref_22","unstructured":"Guo, Z., Zhang, Y., and Lu, W. (August, January 28). Attention Guided Graph Convolutional Networks for Relation Extraction. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Yu, J., Bohnet, B., and Poesio, M. (2020, January 5\u201310). Named Entity Recognition as Dependency Parsing. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online.","DOI":"10.18653\/v1\/2020.acl-main.577"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Nguyen, D.Q., and Verspoor, K. (2019). End-to-end neural relation extraction using deep biaffine attention. Proceedings of the European Conference on Information Retrieval, Springer.","DOI":"10.1007\/978-3-030-15712-8_47"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Luan, Y., He, L., Ostendorf, M., and Hajishirzi, H. (November, January 31). Multi-Task Identification of Entities, Relations, and Coreference for Scientific Knowledge Graph Construction. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium.","DOI":"10.18653\/v1\/D18-1360"},{"key":"ref_26","unstructured":"Ilic, S., Marrese-Taylor, E., Balazs, J., and Matsuo, Y. (November, January 31). Deep contextualized word representations for detecting sarcasm and irony. Proceedings of the 9th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, Brussels, Belgium."},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Wadden, D., Wennberg, U., Luan, Y., and Hajishirzi, H. (2019, January 3\u20137). Entity, Relation, and Event Extraction with Contextualized Span Representations. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China.","DOI":"10.18653\/v1\/D19-1585"},{"key":"ref_28","unstructured":"Eberts, M., and Ulges, A. (September, January 29). Span-based Joint Entity and Relation Extraction with Transformer Pre-training. Proceedings of the 24th European Conference on Artificial Intelligence, Santiago De Compostela, Spain."},{"key":"ref_29","unstructured":"Wu, Y., Schuster, M., Chen, Z., Le, Q.V., and Norouzi, M. (2016). Google\u2019s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation. arXiv."},{"key":"ref_30","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., and Polosukhin, I. (2017, January 4\u20139). Attention is all you need. Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA."},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Ji, B., Yu, J., Li, S., Ma, J., Wu, Q., Tan, Y., and Liu, H. (2020, January 8\u201313). Span-based joint entity and relation extraction with attention-based span-specific and contextual semantic representations. Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain.","DOI":"10.18653\/v1\/2020.coling-main.8"},{"key":"ref_32","unstructured":"Strauss, B., Toma, B., Ritter, A., De Marneffe, M.C., and Xu, W. (2016, January 11\u201312). Results of the wnut16 named entity recognition shared task. Proceedings of the 2nd Workshop on Noisy User-generated Text (WNUT), Osaka, Japan."},{"key":"ref_33","unstructured":"Roth, D., and Yih, W.t. (2019, January 3\u20137). A Linear Programming Formulation for Global Inference in Natural Language Tasks. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Nie, Y., Tian, Y., Wan, X., Song, Y., and Dai, B. (2020, January 16\u201320). Named Entity Recognition for Social Media Texts with Semantic Augmentation. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online.","DOI":"10.18653\/v1\/2020.emnlp-main.107"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Beltagy, I., Lo, K., and Cohan, A. (2019, January 3\u20137). SciBERT: A Pretrained Language Model for Scientific Text. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China.","DOI":"10.18653\/v1\/D19-1371"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Nguyen, D.Q., Vu, T., and Nguyen, A.T. (2020, January 16\u201320). BERTweet: A pre-trained language model for English Tweets. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Online.","DOI":"10.18653\/v1\/2020.emnlp-demos.2"},{"key":"ref_37","unstructured":"Zhou, J.T., Zhang, H., Jin, D., Zhu, H., Fang, M., Goh, R.S.M., and Kwok, K. (August, January 28). Dual adversarial neural transfer for low-resource named entity recognition. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy."},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Shahzad, M., Amin, A., Esteves, D., and Ngomo, A.C.N. (2021, January 17\u201319). InferNER: An attentive model leveraging the sentence-level information for Named Entity Recognition in Microblogs. Proceedings of the The International FLAIRS Conference Proceedings, North Miami Beach, FL, USA.","DOI":"10.32473\/flairs.v34i1.128538"},{"key":"ref_39","doi-asserted-by":"crossref","unstructured":"Wang, X., Jiang, Y., Bach, N., Wang, T., Huang, Z., Huang, F., and Tu, K. (2021, January 1\u20136). Improving Named Entity Recognition by External Context Retrieving and Cooperative Learning. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online.","DOI":"10.18653\/v1\/2021.acl-long.142"},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Yan, Z., Zhang, C., Fu, J., Zhang, Q., and Wei, Z. (2021, January 7\u201311). A Partition Filter Network for Joint Entity and Relation Extraction. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Online.","DOI":"10.18653\/v1\/2021.emnlp-main.17"},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Zhong, Z., and Chen, D. (2021, January 6\u201311). A Frustratingly Easy Approach for Entity and Relation Extraction. Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online.","DOI":"10.18653\/v1\/2021.naacl-main.5"},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"Yu, H., Mao, X.L., Chi, Z., Wei, W., and Huang, H. (2020, January 9\u201311). A robust and domain-adaptive approach for low-resource named entity recognition. Proceedings of the 2020 IEEE International Conference on Knowledge Graph (ICKG), Nanjing, China.","DOI":"10.1109\/ICBK50248.2020.00050"},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"34","DOI":"10.1016\/j.eswa.2018.07.032","article-title":"Joint entity recognition and relation extraction as a multi-head selection problem","volume":"114","author":"Bekoulis","year":"2018","journal-title":"Expert Syst. Appl."},{"key":"ref_44","unstructured":"Tran, T., and Kavuluru, R. (2019). Neural metric learning for fast end-to-end relation extraction. arXiv."},{"key":"ref_45","doi-asserted-by":"crossref","unstructured":"Zhang, M., Zhang, Y., and Fu, G. (2017, January 7\u201311). End-to-end neural relation extraction with global optimization. Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark.","DOI":"10.18653\/v1\/D17-1182"},{"key":"ref_46","unstructured":"Li, X., Yin, F., Sun, Z., Li, X., Yuan, A., Chai, D., Zhou, M., and Li, J. (August, January 28). Entity-Relation Extraction as Multi-Turn Question Answering. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy."},{"key":"ref_47","unstructured":"Crone, P. (2020). Deeper Task-Specificity Improves Joint Entity and Relation Extraction. arXiv."},{"key":"ref_48","doi-asserted-by":"crossref","unstructured":"Wang, J., and Lu, W. (2020, January 16\u201320). Two are Better than One: Joint Entity and Relation Extraction with Table-Sequence Encoders. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online.","DOI":"10.18653\/v1\/2020.emnlp-main.133"}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/8\/2852\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T22:50:10Z","timestamp":1760136610000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/8\/2852"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,8]]},"references-count":48,"journal-issue":{"issue":"8","published-online":{"date-parts":[[2022,4]]}},"alternative-id":["s22082852"],"URL":"https:\/\/doi.org\/10.3390\/s22082852","relation":{},"ISSN":["1424-8220"],"issn-type":[{"value":"1424-8220","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,8]]}}}