{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,7]],"date-time":"2025-12-07T05:14:09Z","timestamp":1765084449029,"version":"3.46.0"},"reference-count":37,"publisher":"MDPI AG","issue":"4","license":[{"start":{"date-parts":[[2025,12,5]],"date-time":"2025-12-05T00:00:00Z","timestamp":1764892800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100012456","name":"National Social Science Foundation of China","doi-asserted-by":"crossref","award":["20BMZ092"],"award-info":[{"award-number":["20BMZ092"]}],"id":[{"id":"10.13039\/501100012456","id-type":"DOI","asserted-by":"crossref"}]},{"name":"SICNU Foundation for Experimental Technology","award":["SYJS2023013"],"award-info":[{"award-number":["SYJS2023013"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Informatics"],"abstract":"<jats:p>With the rapid development of information technology, there is an increasing demand for the digital preservation of traditional festival culture and the extraction of relevant knowledge. However, existing research on Named Entity Recognition (NER) for Chinese traditional festival culture lacks support from high-quality corpora and dedicated model methods. To address this gap, this study proposes a Named Entity Recognition model, CLFF-NER, which integrates multi-source heterogeneous information. The model operates as follows: first, Multilingual BERT is employed to obtain the contextual semantic representations of Chinese and English sentences. Subsequently, a Multiconvolutional Kernel Network (MKN) is used to extract the local structural features of entities. Then, a Transformer module is introduced to achieve cross-lingual, cross-attention fusion of Chinese and English semantics. Furthermore, a Graph Neural Network (GNN) is utilized to selectively supplement useful English information, thereby alleviating the interference caused by redundant information. Finally, a gating mechanism and Conditional Random Field (CRF) are combined to jointly optimize the recognition results. Experiments were conducted on the public Chinese Festival Culture Dataset (CTFCDataSet), and the model achieved 89.45%, 90.01%, and 89.73% in precision, recall, and F1 score, respectively\u2014significantly outperforming a range of mainstream baseline models. Meanwhile, the model also demonstrated competitive performance on two other public datasets, Resume and Weibo, which verifies its strong cross-domain generalization ability.<\/jats:p>","DOI":"10.3390\/informatics12040136","type":"journal-article","created":{"date-parts":[[2025,12,5]],"date-time":"2025-12-05T15:11:24Z","timestamp":1764947484000},"page":"136","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["CLFF-NER: A Cross-Lingual Feature Fusion Model for Named Entity Recognition in the Traditional Chinese Festival Culture Domain"],"prefix":"10.3390","volume":"12","author":[{"given":"Shenghe","family":"Yang","sequence":"first","affiliation":[{"name":"School of Computer Science and Technology, Changchun Normal University, Changchun 130000, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-3500-2276","authenticated-orcid":false,"given":"Kun","family":"He","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, Changchun Normal University, Changchun 130000, China"}]},{"given":"Wei","family":"Li","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, Sichuan Normal University, Chengdu 610000, China"}]},{"given":"Yingying","family":"He","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, Changchun Normal University, Changchun 130000, China"}]}],"member":"1968","published-online":{"date-parts":[[2025,12,5]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"103352","DOI":"10.1016\/j.ipm.2023.103352","article-title":"Planarized sentence representation for nested named entity recognition","volume":"60","author":"Geng","year":"2023","journal-title":"Inf. Process. Manag."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"121925","DOI":"10.1016\/j.eswa.2023.121925","article-title":"Chinese engineering geological named entity recognition by fusing multi-features and data enhancement using deep learning","volume":"238","author":"Qiu","year":"2024","journal-title":"Expert Syst. Appl."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"111266","DOI":"10.1016\/j.knosys.2023.111266","article-title":"Exploiting global contextual information for document-level named entity recognition","volume":"284","author":"Yu","year":"2024","journal-title":"Knowl.-Based Syst."},{"key":"ref_4","unstructured":"Devlin, J., Chang, M.W., Lee, K., and Toutanova, K. (2018). BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv."},{"key":"ref_5","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., and Soricut, R. (2020). ALBERT: A lite BERT for self-supervised learning of language representations. arXiv."},{"key":"ref_6","unstructured":"Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., and Le, Q.V. (2019). XLNet: Generalized autoregressive pretraining for language understanding. arXiv."},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Hu, S., Zhang, H., Hu, X., and Du, J. (2022, January 26\u201328). Chinese Named Entity Recognition based on BERT-CRF Model. Proceedings of the 2022 IEEE\/ACIS 22nd International Conference on Computer and Information Science (ICIS), Zhuhai, China.","DOI":"10.1109\/ICIS54925.2022.9882514"},{"key":"ref_8","unstructured":"Shen, Y., Tan, M., Sordoni, A., and Courville, A. (2018). Ordered neurons: Integrating tree structures into recurrent neural networks. arXiv."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"121103","DOI":"10.1016\/j.eswa.2023.121103","article-title":"Naming entity recognition of citrus pests and diseases based on the BERT-BiLSTM-CRF model","volume":"234","author":"Liu","year":"2023","journal-title":"Expert Syst. Appl."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"103767","DOI":"10.1016\/j.inffus.2025.103767","article-title":"Multimodal named entity recognition in the era of large pre-trained models: A comprehensive survey","volume":"127","author":"Xu","year":"2026","journal-title":"Inf. Fusion"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"108106","DOI":"10.1016\/j.neunet.2025.108106","article-title":"MINIGE-MNER: A multi-stage interaction network inspired by gene editing for multimodal named entity recognition","volume":"194","author":"Kong","year":"2026","journal-title":"Neural Netw."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"103405","DOI":"10.1016\/j.inffus.2025.103405","article-title":"Multimodal Named Entity Recognition based on topic prompt and multi-curriculum denoising","volume":"124","author":"Xu","year":"2025","journal-title":"Inf. Fusion"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"110649","DOI":"10.1016\/j.engappai.2025.110649","article-title":"A Chinese medical named entity recognition method considering length diversity of entities","volume":"150","author":"Zhang","year":"2025","journal-title":"Eng. Appl. Artif. Intell."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"50","DOI":"10.1109\/TKDE.2020.2981314","article-title":"A Survey on Deep Learning for Named Entity Recognition","volume":"34","author":"Li","year":"2022","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Francis, S., Van Landeghem, J., and Moens, M.-F. (2019). Transfer Learning for Named Entity Recognition in Financial and Biomedical Documents. Information, 10.","DOI":"10.3390\/info10080248"},{"key":"ref_16","unstructured":"Hachey, B., and Webster, K. (2015). Domain Adaption of Named Entity Recognition to Support Credit Risk Assessment. Proceedings of Australasian Language Technology Association Workshop, ACL. Available online: https:\/\/aclanthology.org\/U15-1010\/."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Song, B., Li, F., Liu, Y., and Zeng, X. (2021). Deep learning methods for biomedical named entity recognition: A survey and qualitative comparison. Brief Bioinform., 22.","DOI":"10.1093\/bib\/bbab282"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Loukas, L., Fergadiotis, M., Chalkidis, I., Spyropoulou, E., Malakasiotis, P., Androutsopoulos, I., and Paliouras, G. (2022). FiNER: Financial Numeric Entity Recognition for XBRL Tagging. arXiv.","DOI":"10.18653\/v1\/2022.acl-long.303"},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Wang, N., Yang, H., and Wang, C.D. (2023). FinGPT: Instruction Tuning Benchmark for Open-Source Large Language Models in Financial Datasets. arXiv.","DOI":"10.2139\/ssrn.4489826"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Ogrinc, M., Korou\u0161i\u0107 Seljak, B., and Eftimov, T. (2024). Zero-shot evaluation of ChatGPT for food named-entity recognition and linking. Front. Nutr., 11.","DOI":"10.3389\/fnut.2024.1429259"},{"key":"ref_21","unstructured":"Mihova, M., and Jovanov, M. (2024). Extracting Entities and Relations in Analyst Stock Ratings News. ICT Innovations 2023. Learning: Humans, Theory, Machines, and Data, Springer. Communications in Computer and Information Science."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"107389","DOI":"10.1016\/j.neunet.2025.107389","article-title":"A shape composition method for named entity recognition","volume":"187","author":"Hu","year":"2025","journal-title":"Neural Netw."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Katz, D.M., Hartung, D., Gerlach, L., Jana, A., and Bommarito, M.J. (2023). Natural language processing in the legal domain. arXiv.","DOI":"10.2139\/ssrn.4336224"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Lee, K., Wei, C.H., and Lu, Z. (2021). Recent advances of automated methods for searching and extracting genomic variant information from biomedical literature. Brief Bioinform., 22.","DOI":"10.1093\/bib\/bbaa142"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"100511","DOI":"10.1016\/j.cosrev.2022.100511","article-title":"Neural natural language processing for unstructured data in electronic health records: A review","volume":"46","author":"Li","year":"2022","journal-title":"Comput. Sci. Rev."},{"key":"ref_26","first-page":"2493","article-title":"Natural language processing (almost) from scratch","volume":"12","author":"Collobert","year":"2011","journal-title":"J. Mach. Learn. Res."},{"key":"ref_27","unstructured":"Daelemans, W., and Osborne, M. (June, January 31). Named entity recognition with long short-term memory. Proceedings of the Seventh Conference on Natural Language Learning, coNLL 2003, Edmonton, AB, Canada. Available online: https:\/\/aclanthology.org\/W03-0426\/."},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Lample, G., Ballesteros, M., Subramanian, S., Kawakami, K., and Dyer, C. (2016). Neural architectures for named entity recognition. arXiv.","DOI":"10.18653\/v1\/N16-1030"},{"key":"ref_29","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., and Polosukhin, I. (2023). Attention Is All You Need. arXiv."},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"112158","DOI":"10.1016\/j.asoc.2024.112158","article-title":"Evaluating Named Entity Recognition: A comparative analysis of mono- and multilingual transformer models on a novel Brazilian corporate earnings call transcripts dataset","volume":"166","author":"Abilio","year":"2024","journal-title":"Appl. Soft Comput."},{"key":"ref_31","unstructured":"Huang, Z., Xu, W., and Yu, K. (2015). Bidirectional LSTM-CRF Models for Sequence Tagging. arXiv."},{"key":"ref_32","unstructured":"Mao, Q., Li, J., and Meng, K. (2022). Improving Chinese Named Entity Recognition by Search Engine Augmentation. arXiv."},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Gu, R., Wang, T., Deng, J., and Cheng, L. (2023). Improving Chinese Named Entity Recognition by Interactive Fusion of Contextual Representation and Glyph Representation. Appl. Sci., 13.","DOI":"10.3390\/app13074299"},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Hu, W., He, L., Ma, H., Wang, K., and Xiao, J. (2022). KGNER: Improving Chinese Named Entity Recognition by BERT Infused with the Knowledge Graph. Appl. Sci., 12.","DOI":"10.3390\/app12157702"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Fang, Q., Li, Y., Feng, H., and Ruan, Y. (2023). Chinese Named Entity Recognition Model Based on Multi-Task Learning. Appl. Sci., 13.","DOI":"10.3390\/app13084770"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Sun, S., Deng, M., Yu, X., and Zhao, L. (2025). HREB-CRF: Hierarchical Reduced-bias EMA for Chinese Named Entity Recognition. arXiv.","DOI":"10.1109\/IJCNN64981.2025.11227588"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Li, H., Cheng, M., Yang, Z., Yang, L., and Chua, Y. (2023). Named entity recognition for Chinese based on global pointer and adversarial training. Sci. Rep., 13.","DOI":"10.1038\/s41598-023-30355-y"}],"container-title":["Informatics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2227-9709\/12\/4\/136\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,7]],"date-time":"2025-12-07T05:13:36Z","timestamp":1765084416000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2227-9709\/12\/4\/136"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,5]]},"references-count":37,"journal-issue":{"issue":"4","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["informatics12040136"],"URL":"https:\/\/doi.org\/10.3390\/informatics12040136","relation":{},"ISSN":["2227-9709"],"issn-type":[{"type":"electronic","value":"2227-9709"}],"subject":[],"published":{"date-parts":[[2025,12,5]]}}}