{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T03:56:08Z","timestamp":1742961368700,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":49,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819608393"},{"type":"electronic","value":"9789819608409"}],"license":[{"start":{"date-parts":[[2024,12,13]],"date-time":"2024-12-13T00:00:00Z","timestamp":1734048000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,13]],"date-time":"2024-12-13T00:00:00Z","timestamp":1734048000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-0840-9_15","type":"book-chapter","created":{"date-parts":[[2024,12,12]],"date-time":"2024-12-12T17:28:55Z","timestamp":1734024535000},"page":"215-229","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["GKF-mQA: Generative Knowledge Fusion Based on\u00a0Large Language Models for\u00a0Enhancing Medical Question Answering"],"prefix":"10.1007","author":[{"given":"Xinbai","family":"Li","sequence":"first","affiliation":[]},{"given":"Man","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,13]]},"reference":[{"key":"15_CR1","doi-asserted-by":"crossref","unstructured":"Alzubi, J.A., Jain, R., Singh, A., Parwekar, P., Gupta, M.: Cobert: Covid-19 question answering system using bert. Arabian journal for science and engineering pp. 1\u201311 (2021)","DOI":"10.1007\/s13369-021-05810-5"},{"key":"15_CR2","doi-asserted-by":"crossref","unstructured":"Bodenreider, O.: The unified medical language system (umls): integrating biomedical terminology. Nucleic acids research 32(suppl_1), D267\u2013D270 (2004)","DOI":"10.1093\/nar\/gkh061"},{"key":"15_CR3","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"10","key":"15_CR4","first-page":"1","volume":"33","author":"O Byambasuren","year":"2019","unstructured":"Byambasuren, O., Yang, Y., Sui, Z., Dai, D., Chang, B., Li, S., Zan, H.: Preliminary study on the construction of chinese medical knowledge graph. Journal of Chinese Information Processing 33(10), 1\u20139 (2019)","journal-title":"Journal of Chinese Information Processing"},{"key":"15_CR5","doi-asserted-by":"publisher","unstructured":"Cui, L., Wu, Y., Liu, S., Zhang, Y.: Knowledge enhanced fine-tuning for better handling unseen entities in dialogue generation. In: Moens, M.F., Huang, X., Specia, L., Yih, S.W.t. (eds.) Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. pp. 2328\u20132337. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic (Nov 2021). https:\/\/doi.org\/10.18653\/v1\/2021.emnlp-main.179, https:\/\/aclanthology.org\/2021.emnlp-main.179","DOI":"10.18653\/v1\/2021.emnlp-main.179"},{"issue":"5","key":"15_CR6","doi-asserted-by":"publisher","first-page":"760","DOI":"10.1016\/j.jbi.2009.08.007","volume":"42","author":"D Demner-Fushman","year":"2009","unstructured":"Demner-Fushman, D., Chapman, W.W., McDonald, C.J.: What can natural language processing do for clinical decision support? J. Biomed. Inform. 42(5), 760\u2013772 (2009)","journal-title":"J. Biomed. Inform."},{"key":"15_CR7","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"15_CR8","doi-asserted-by":"crossref","unstructured":"Hayashi, H., Hu, Z., Xiong, C., Neubig, G.: Latent relation language models. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol.\u00a034, pp. 7911\u20137918 (2020)","DOI":"10.1609\/aaai.v34i05.6298"},{"key":"15_CR9","doi-asserted-by":"crossref","unstructured":"He, L., Zheng, S., Yang, T., Zhang, F.: Klmo: Knowledge graph enhanced pretrained language model with fine-grained relationships. In: Findings of the Association for Computational Linguistics: EMNLP 2021. pp. 4536\u20134542 (2021)","DOI":"10.18653\/v1\/2021.findings-emnlp.384"},{"key":"15_CR10","doi-asserted-by":"publisher","unstructured":"He, Y., Zhu, Z., Zhang, Y., Chen, Q., Caverlee, J.: Infusing Disease Knowledge into BERT for Health Question Answering, Medical Inference and Disease Name Recognition. In: Webber, B., Cohn, T., He, Y., Liu, Y. (eds.) Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). pp. 4604\u20134614. Association for Computational Linguistics, Online (Nov 2020).https:\/\/doi.org\/10.18653\/v1\/2020.emnlp-main.372, https:\/\/aclanthology.org\/2020.emnlp-main.372","DOI":"10.18653\/v1\/2020.emnlp-main.372"},{"key":"15_CR11","doi-asserted-by":"crossref","unstructured":"He, Y., Zhu, Z., Zhang, Y., Chen, Q., Caverlee, J.: Infusing disease knowledge into bert for health question answering, medical inference and disease name recognition. arXiv preprint arXiv:2010.03746 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.372"},{"issue":"17","key":"15_CR12","doi-asserted-by":"publisher","first-page":"23777","DOI":"10.1007\/s11042-022-12820-4","volume":"81","author":"IC Hsu","year":"2022","unstructured":"Hsu, I.C., Yu, J.D.: A medical chatbot using machine learning and natural language understanding. Multimedia Tools and Applications 81(17), 23777\u201323799 (2022)","journal-title":"Multimedia Tools and Applications"},{"key":"15_CR13","unstructured":"Huang, K., Altosaar, J., Ranganath, R.: Clinicalbert: Modeling clinical notes and predicting hospital readmission. arXiv preprint arXiv:1904.05342 (2019)"},{"key":"15_CR14","doi-asserted-by":"publisher","unstructured":"Jiang, J., Wang, S., Li, Q., Kong, L., Wu, C.: A cognitive stimulation dialogue system with multi-source knowledge fusion for elders with cognitive impairment. In: Rogers, A., Boyd-Graber, J., Okazaki, N. (eds.) Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). pp. 10628\u201310640. Association for Computational Linguistics, Toronto, Canada (Jul 2023https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.593, https:\/\/aclanthology.org\/2023.acl-long.593","DOI":"10.18653\/v1\/2023.acl-long.593"},{"issue":"14","key":"15_CR15","doi-asserted-by":"publisher","first-page":"6421","DOI":"10.3390\/app11146421","volume":"11","author":"D Jin","year":"2021","unstructured":"Jin, D., Pan, E., Oufattole, N., Weng, W.H., Fang, H., Szolovits, P.: What disease does this patient have? a large-scale open domain question answering dataset from medical exams. Appl. Sci. 11(14), 6421 (2021)","journal-title":"Appl. Sci."},{"key":"15_CR16","doi-asserted-by":"publisher","unstructured":"Lee, J., Yoon, W., Kim, S., Kim, D., Kim, S., So, C.H., Kang, J.: BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36(4), 1234\u20131240 (09 2019).https:\/\/doi.org\/10.1093\/bioinformatics\/btz682, https:\/\/doi.org\/10.1093\/bioinformatics\/btz682","DOI":"10.1093\/bioinformatics\/btz682"},{"issue":"12","key":"15_CR17","doi-asserted-by":"publisher","first-page":"191","DOI":"10.3390\/a11120191","volume":"11","author":"C Li","year":"2018","unstructured":"Li, C., Annisa, A., Zaman, A., Qaosar, M., Ahmed, S., Morimoto, Y.: Mapreduce algorithm for location recommendation by using area skyline query. Algorithms 11(12), 191 (2018)","journal-title":"Algorithms"},{"key":"15_CR18","doi-asserted-by":"crossref","unstructured":"Li, C., Yamanishi, Y.: Gxvaes: Two joint vaes generate hit molecules from gene expression profiles. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol.\u00a038, pp. 13455\u201313463 (2024)","DOI":"10.1609\/aaai.v38i12.29248"},{"key":"15_CR19","unstructured":"Li, C., Yamanishi, Y.: Tengan: Pure transformer encoders make an efficient discrete gan for de novo molecular generation. In: International Conference on Artificial Intelligence and Statistics. pp. 361\u2013369. PMLR (2024)"},{"key":"15_CR20","doi-asserted-by":"crossref","unstructured":"Liu, W., Zhou, P., Zhao, Z., Wang, Z., Ju, Q., Deng, H., Wang, P.: K-bert: Enabling language representation with knowledge graph. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol.\u00a034, pp. 2901\u20132908 (2020)","DOI":"10.1609\/aaai.v34i03.5681"},{"key":"15_CR21","doi-asserted-by":"publisher","first-page":"2320","DOI":"10.3389\/fpubh.2021.831404","volume":"9","author":"V Mani","year":"2022","unstructured":"Mani, V., Kavitha, C., Band, S.S., Mosavi, A., Hollins, P., Palanisamy, S.: A recommendation system based on ai for storing block data in the electronic health repository. Front. Public Health 9, 2320 (2022)","journal-title":"Front. Public Health"},{"key":"15_CR22","doi-asserted-by":"publisher","unstructured":"Neumann, M., King, D., Beltagy, I., Ammar, W.: ScispaCy: Fast and Robust Models for Biomedical Natural Language Processing. In: Proceedings of the 18th BioNLP Workshop and Shared Task. pp. 319\u2013327. Association for Computational Linguistics, Florence, Italy (Aug 2019).https:\/\/doi.org\/10.18653\/v1\/W19-5034, https:\/\/www.aclweb.org\/anthology\/W19-5034","DOI":"10.18653\/v1\/W19-5034"},{"key":"15_CR23","doi-asserted-by":"crossref","unstructured":"Pampari, A., Raghavan, P., Liang, J., Peng, J.: emrqa: A large corpus for question answering on electronic medical records. arXiv preprint arXiv:1809.00732 (2018)","DOI":"10.18653\/v1\/D18-1258"},{"key":"15_CR24","doi-asserted-by":"crossref","unstructured":"Peng, Y., Yan, S., Lu, Z.: Transfer learning in biomedical natural language processing: an evaluation of bert and elmo on ten benchmarking datasets. arXiv preprint arXiv:1906.05474 (2019)","DOI":"10.18653\/v1\/W19-5006"},{"key":"15_CR25","doi-asserted-by":"publisher","unstructured":"Peters, M.E., Neumann, M., Logan, R., Schwartz, R., Joshi, V., Singh, S., Smith, N.A.: Knowledge enhanced contextual word representations. In: Inui, K., Jiang, J., Ng, V., Wan, X. (eds.) Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). pp. 43\u201354. Association for Computational Linguistics, Hong Kong, China (Nov 2019).https:\/\/doi.org\/10.18653\/v1\/D19-1005, https:\/\/aclanthology.org\/D19-1005","DOI":"10.18653\/v1\/D19-1005"},{"issue":"10","key":"15_CR26","doi-asserted-by":"publisher","first-page":"1872","DOI":"10.1007\/s11431-020-1647-3","volume":"63","author":"X Qiu","year":"2020","unstructured":"Qiu, X., Sun, T., Xu, Y., Shao, Y., Dai, N., Huang, X.: Pre-trained models for natural language processing: A survey. SCIENCE CHINA Technol. Sci. 63(10), 1872\u20131897 (2020)","journal-title":"SCIENCE CHINA Technol. Sci."},{"issue":"1","key":"15_CR27","first-page":"5485","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., Liu, P.J.: Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research 21(1), 5485\u20135551 (2020)","journal-title":"The Journal of Machine Learning Research"},{"key":"15_CR28","doi-asserted-by":"crossref","unstructured":"Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: Squad: 100,000+ questions for machine comprehension of text. arXiv preprint arXiv:1606.05250 (2016)","DOI":"10.18653\/v1\/D16-1264"},{"key":"15_CR29","doi-asserted-by":"publisher","unstructured":"Rawat, B.P.S., Weng, W.H., Min, S.Y., Raghavan, P., Szolovits, P.: Entity-enriched neural models for clinical question answering. In: Demner-Fushman, D., Cohen, K.B., Ananiadou, S., Tsujii, J. (eds.) Proceedings of the 19th SIGBioMed Workshop on Biomedical Language Processing. pp. 112\u2013122. Association for Computational Linguistics, Online (Jul 2020).https:\/\/doi.org\/10.18653\/v1\/2020.bionlp-1.12, https:\/\/aclanthology.org\/2020.bionlp-1.12","DOI":"10.18653\/v1\/2020.bionlp-1.12"},{"key":"15_CR30","doi-asserted-by":"crossref","unstructured":"Rawat, B.P.S., Weng, W.H., Min, S.Y., Raghavan, P., Szolovits, P.: Entity-enriched neural models for clinical question answering. In: Proceedings of the 19th SIGBioMed Workshop on Biomedical Language Processing. pp. 112\u2013122 (2020)","DOI":"10.18653\/v1\/2020.bionlp-1.12"},{"key":"15_CR31","doi-asserted-by":"publisher","unstructured":"Remy, F., Demuynck, K., Demeester, T.: BioLORD: Learning ontological representations from definitions for biomedical concepts and their textual descriptions. In: Goldberg, Y., Kozareva, Z., Zhang, Y. (eds.) Findings of the Association for Computational Linguistics: EMNLP 2022. pp. 1454\u20131465. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates (Dec 2022https:\/\/doi.org\/10.18653\/v1\/2022.findings-emnlp.104, https:\/\/aclanthology.org\/2022.findings-emnlp.104","DOI":"10.18653\/v1\/2022.findings-emnlp.104"},{"key":"15_CR32","doi-asserted-by":"publisher","first-page":"842","DOI":"10.1162\/tacl_a_00349","volume":"8","author":"A Rogers","year":"2021","unstructured":"Rogers, A., Kovaleva, O., Rumshisky, A.: A primer in bertology: What we know about how bert works. Transactions of the Association for Computational Linguistics 8, 842\u2013866 (2021)","journal-title":"Transactions of the Association for Computational Linguistics"},{"key":"15_CR33","unstructured":"Su, J.: T5 pegasus - zhuiyiai. Tech. rep. (2021), https:\/\/github.com\/ZhuiyiTechnology\/t5-pegasus"},{"key":"15_CR34","doi-asserted-by":"publisher","unstructured":"Sun, T., Shao, Y., Qiu, X., Guo, Q., Hu, Y., Huang, X., Zhang, Z.: CoLAKE: Contextualized language and knowledge embedding. In: Scott, D., Bel, N., Zong, C. (eds.) Proceedings of the 28th International Conference on Computational Linguistics. pp. 3660\u20133670. International Committee on Computational Linguistics, Barcelona, Spain (Online) (Dec 2020).https:\/\/doi.org\/10.18653\/v1\/2020.coling-main.327, https:\/\/aclanthology.org\/2020.coling-main.327","DOI":"10.18653\/v1\/2020.coling-main.327"},{"key":"15_CR35","unstructured":"Sun, Y., Wang, S., Li, Y., Feng, S., Chen, X., Zhang, H., Tian, X., Zhu, D., Tian, H., Wu, H.: Ernie: Enhanced representation through knowledge integration. arXiv preprint arXiv:1904.09223 (2019)"},{"issue":"1","key":"15_CR36","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1038\/s41746-020-0221-y","volume":"3","author":"RT Sutton","year":"2020","unstructured":"Sutton, R.T., Pincock, D., Baumgart, D.C., Sadowski, D.C., Fedorak, R.N., Kroeker, K.I.: An overview of clinical decision support systems: benefits, risks, and strategies for success. NPJ digital medicine 3(1), 17 (2020)","journal-title":"NPJ digital medicine"},{"key":"15_CR37","doi-asserted-by":"publisher","unstructured":"Tang, S., Chai, H., Yao, Z., Ding, Y., Gao, C., Fang, B., Liao, Q.: Affective knowledge enhanced multiple-graph fusion networks for aspect-based sentiment analysis. In: Goldberg, Y., Kozareva, Z., Zhang, Y. (eds.) Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing. pp. 5352\u20135362. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates (Dec 2022https:\/\/doi.org\/10.18653\/v1\/2022.emnlp-main.359, https:\/\/aclanthology.org\/2022.emnlp-main.359","DOI":"10.18653\/v1\/2022.emnlp-main.359"},{"key":"15_CR38","unstructured":"Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., et\u00a0al.: Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)"},{"key":"15_CR39","doi-asserted-by":"publisher","unstructured":"Wang, J., Huang, W., Qiu, M., Shi, Q., Wang, H., Li, X., Gao, M.: Knowledge prompting in pre-trained language model for natural language understanding. In: Goldberg, Y., Kozareva, Z., Zhang, Y. (eds.) Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing. pp. 3164\u20133177. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates (Dec 2022https:\/\/doi.org\/10.18653\/v1\/2022.emnlp-main.207, https:\/\/aclanthology.org\/2022.emnlp-main.207","DOI":"10.18653\/v1\/2022.emnlp-main.207"},{"key":"15_CR40","doi-asserted-by":"crossref","unstructured":"Wang, R., Tang, D., Duan, N., Wei, Z., Huang, X., Cao, G., Jiang, D., Zhou, M., et\u00a0al.: K-adapter: Infusing knowledge into pre-trained models with adapters. arXiv preprint arXiv:2002.01808 (2020)","DOI":"10.18653\/v1\/2021.findings-acl.121"},{"key":"15_CR41","doi-asserted-by":"publisher","first-page":"176","DOI":"10.1162\/tacl_a_00360","volume":"9","author":"X Wang","year":"2021","unstructured":"Wang, X., Gao, T., Zhu, Z., Zhang, Z., Liu, Z., Li, J., Tang, J.: Kepler: A unified model for knowledge embedding and pre-trained language representation. Transactions of the Association for Computational Linguistics 9, 176\u2013194 (2021)","journal-title":"Transactions of the Association for Computational Linguistics"},{"key":"15_CR42","doi-asserted-by":"publisher","unstructured":"Wenrong, L.: MedicalQA:Medical dataset for machine reading comprehension (Aug 2022https:\/\/doi.org\/10.11922\/sciencedb.j00001.00411","DOI":"10.11922\/sciencedb.j00001.00411"},{"key":"15_CR43","doi-asserted-by":"publisher","unstructured":"Xia, F., Li, B., Weng, Y., He, S., Liu, K., Sun, B., Li, S., Zhao, J.: MedConQA: Medical conversational question answering system based on knowledge graphs. In: Che, W., Shutova, E. (eds.) Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. pp. 148\u2013158. Association for Computational Linguistics, Abu Dhabi, UAE (Dec 2022https:\/\/doi.org\/10.18653\/v1\/2022.emnlp-demos.15, https:\/\/aclanthology.org\/2022.emnlp-demos.15","DOI":"10.18653\/v1\/2022.emnlp-demos.15"},{"key":"15_CR44","unstructured":"Xiong, W., Du, J., Wang, W.Y., Stoyanov, V.: Pretrained encyclopedia: Weakly supervised knowledge-pretrained language model. arXiv preprint arXiv:1912.09637 (2019)"},{"key":"15_CR45","doi-asserted-by":"publisher","unstructured":"Yang, H., Huang, S., Dai, X.Y., Chen, J.: Fine-grained knowledge fusion for sequence labeling domain adaptation. In: Inui, K., Jiang, J., Ng, V., Wan, X. (eds.) Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). pp. 4197\u20134206. Association for Computational Linguistics, Hong Kong, China (Nov 2019).https:\/\/doi.org\/10.18653\/v1\/D19-1429, https:\/\/aclanthology.org\/D19-1429","DOI":"10.18653\/v1\/D19-1429"},{"key":"15_CR46","doi-asserted-by":"publisher","unstructured":"Yu, D., Zhu, C., Fang, Y., Yu, W., Wang, S., Xu, Y., Ren, X., Yang, Y., Zeng, M.: KG-FiD: Infusing knowledge graph in fusion-in-decoder for open-domain question answering. In: Muresan, S., Nakov, P., Villavicencio, A. (eds.) Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). pp. 4961\u20134974. Association for Computational Linguistics, Dublin, Ireland (May 2022).https:\/\/doi.org\/10.18653\/v1\/2022.acl-long.340, https:\/\/aclanthology.org\/2022.acl-long.340","DOI":"10.18653\/v1\/2022.acl-long.340"},{"key":"15_CR47","doi-asserted-by":"crossref","unstructured":"Zhai, J., Zheng, X., Wang, C.D., Li, H., Tian, Y.: Knowledge prompt-tuning for sequential recommendation. In: Proceedings of the 31st ACM International Conference on Multimedia. pp. 6451\u20136461 (2023)","DOI":"10.1145\/3581783.3612252"},{"key":"15_CR48","doi-asserted-by":"publisher","unstructured":"Zhang, Z., Han, X., Liu, Z., Jiang, X., Sun, M., Liu, Q.: ERNIE: Enhanced language representation with informative entities. In: Korhonen, A., Traum, D., M\u00e0rquez, L. (eds.) Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. pp. 1441\u20131451. Association for Computational Linguistics, Florence, Italy (Jul 2019).https:\/\/doi.org\/10.18653\/v1\/P19-1139, https:\/\/aclanthology.org\/P19-1139","DOI":"10.18653\/v1\/P19-1139"},{"key":"15_CR49","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2023.126488","volume":"552","author":"Z Zhu","year":"2023","unstructured":"Zhu, Z., Mao, K.: Knowledge-based bert word embedding fine-tuning for emotion recognition. Neurocomputing 552, 126488 (2023)","journal-title":"Neurocomputing"}],"container-title":["Lecture Notes in Computer Science","Advanced Data Mining and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-0840-9_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,12]],"date-time":"2024-12-12T18:09:01Z","timestamp":1734026941000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-0840-9_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,13]]},"ISBN":["9789819608393","9789819608409"],"references-count":49,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-0840-9_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,12,13]]},"assertion":[{"value":"13 December 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ADMA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Advanced Data Mining and Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Sydney, NSW","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Australia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 December 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 December 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"adma2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/adma2024.github.io\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}