{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T19:31:23Z","timestamp":1769023883820,"version":"3.49.0"},"publisher-location":"Cham","reference-count":43,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031402913","type":"print"},{"value":"9783031402920","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-40292-0_2","type":"book-chapter","created":{"date-parts":[[2023,8,8]],"date-time":"2023-08-08T23:02:48Z","timestamp":1691535768000},"page":"12-24","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Advancing Domain Adaptation of\u00a0BERT by\u00a0Learning Domain Term Semantics"],"prefix":"10.1007","author":[{"given":"Jian","family":"Yang","sequence":"first","affiliation":[]},{"given":"Xinyu","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Weichun","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Hao","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Yulong","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Gang","family":"Xiao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,9]]},"reference":[{"key":"2_CR1","doi-asserted-by":"crossref","unstructured":"Alsentzer, E., et al.: Publicly available clinical BERT embeddings. In: Proceedings of the 2nd Clinical Natural Language Processing Workshop, pp. 72\u201378. Association for Computational Linguistics, Minneapolis, Minnesota, USA (2019)","DOI":"10.18653\/v1\/W19-1909"},{"key":"2_CR2","unstructured":"Beltagy, I., Cohan, A., Lo, K.: SciBERT: pretrained contextualized embeddings for scientific text. CoRR abs\/1903.10676 (2019). http:\/\/arxiv.org\/abs\/1903.10676"},{"key":"2_CR3","doi-asserted-by":"crossref","unstructured":"Bravo, A., Gonz\u00e1lez, J.P., Queralt-Rosinach, N., Rautschka, M., Furlong, L.I.: Extraction of relations between genes and diseases from text and large-scale data analysis: implications for translational research. BMC Bioinform. 16, 1\u201317 (2015)","DOI":"10.1186\/s12859-015-0472-9"},{"key":"2_CR4","unstructured":"Brown, T.B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., et al.: Language models are few-shot learners. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H.T. (eds.) Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS, 6\u201312 December 2020, virtual (2020)"},{"key":"2_CR5","doi-asserted-by":"crossref","unstructured":"Chalkidis, I., Fergadiotis, M., Malakasiotis, P., Aletras, N., Androutsopoulos, I.: LEGAL-BERT: the muppets straight out of law school. arXiv preprint arXiv:2010.02559 (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.261"},{"key":"2_CR6","doi-asserted-by":"crossref","unstructured":"Crichton, G.K.O., Pyysalo, S., Chiu, B., Korhonen, A.: A neural network multi-task learning approach to biomedical named entity recognition. BMC Bioinform. 18(1), 1\u201314 (2017)","DOI":"10.1186\/s12859-017-1776-8"},{"key":"2_CR7","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Burstein, J., Doran, C., Solorio, T. (eds.) Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, vol. 1, pp. 4171\u20134186. Association for Computational Linguistics, Minneapolis, MN, USA (2019)"},{"key":"2_CR8","doi-asserted-by":"crossref","unstructured":"Gu, Y., Tinn, R., Cheng, H., Lucas, M., Usuyama, N., Liu, X., et al.: Domain-specific language model pretraining for biomedical natural language processing. ACM Trans. Comput. Healthcare 3(1), 1\u201323 (2022)","DOI":"10.1145\/3458754"},{"key":"2_CR9","doi-asserted-by":"crossref","unstructured":"Gururangan, S., Marasovi\u0107, A., Swayamdipta, S., Lo, K., Beltagy, I., Downey, D., Smith, N.A.: Don\u2019t stop pretraining: adapt language models to domains and tasks. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 8342\u20138360. Association for Computational Linguistics (2020)","DOI":"10.18653\/v1\/2020.acl-main.740"},{"key":"2_CR10","doi-asserted-by":"crossref","unstructured":"Hadsell, R., Chopra, S., LeCun, Y.: Dimensionality reduction by learning an invariant mapping. In: 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2006), pp. 1735\u20131742. IEEE Computer Society, New York, NY, USA (2006)","DOI":"10.1109\/CVPR.2006.100"},{"issue":"1","key":"2_CR11","doi-asserted-by":"publisher","first-page":"57","DOI":"10.1016\/S0092-8674(00)81683-9","volume":"100","author":"D Hanahan","year":"2000","unstructured":"Hanahan, D., Weinberg, R.A.: The hallmarks of cancer. Cell 100(1), 57\u201370 (2000)","journal-title":"Cell"},{"key":"2_CR12","doi-asserted-by":"crossref","unstructured":"Hao, B., Zhu, H., Paschalidis, I.C.: Enhancing clinical BERT embedding using a biomedical knowledge base. In: Proceedings of the 28th International Conference on Computational Linguistics, COLING 2020, pp. 657\u2013661. International Committee on Computational Linguistics (2020)","DOI":"10.18653\/v1\/2020.coling-main.57"},{"key":"2_CR13","doi-asserted-by":"crossref","unstructured":"He, B., et al.: BERT-MK: integrating graph contextualized knowledge into pre-trained language models. In: Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 2281\u20132290. Association for Computational Linguistics (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.207"},{"issue":"5","key":"2_CR14","doi-asserted-by":"publisher","first-page":"914","DOI":"10.1016\/j.jbi.2013.07.011","volume":"46","author":"M Herrero-Zazo","year":"2013","unstructured":"Herrero-Zazo, M., Segura-Bedmar, I., Mart\u00ednez, P., Declerck, T.: The DDI corpus: an annotated corpus with pharmacological substances and drug-drug interactions. J. Biomed. Inform. 46(5), 914\u2013920 (2013)","journal-title":"J. Biomed. Inform."},{"key":"2_CR15","doi-asserted-by":"crossref","unstructured":"Hong, J., Kim, T., Lim, H., Choo, J.: Avocado: strategy for adapting vocabulary to downstream domain. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 4692\u20134700. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.385"},{"key":"2_CR16","doi-asserted-by":"crossref","unstructured":"Jin, Q., Dhingra, B., Liu, Z., Cohen, W., Lu, X.: PubMedQA: a dataset for biomedical research question answering. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 2567\u20132577. Association for Computational Linguistics, Hong Kong, China (2019)","DOI":"10.18653\/v1\/D19-1259"},{"key":"2_CR17","doi-asserted-by":"crossref","unstructured":"Kalyan, K.S., Rajasekharan, A., Sangeetha, S.: AMMUS: a survey of transformer-based pretrained models in natural language processing. arXiv preprint arXiv:2108.05542 (2021)","DOI":"10.1016\/j.jbi.2021.103982"},{"key":"2_CR18","volume-title":"3rd International Conference on Learning Representations, ICLR 2015","author":"DP Kingma","year":"2015","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Bengio, Y., LeCun, Y. (eds.) 3rd International Conference on Learning Representations, ICLR 2015. San Diego, CA, USA (2015)"},{"key":"2_CR19","doi-asserted-by":"crossref","unstructured":"Koto, F., Lau, J.H., Baldwin, T.: IndoBERTweet: a pretrained language model for Indonesian twitter with effective domain-specific vocabulary initialization. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 10660\u201310668. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.833"},{"key":"2_CR20","unstructured":"Krallinger, M., et al.: Overview of the BioCreative VI chemical-protein interaction track. In: Proceedings of the Sixth BioCreative Challenge Evaluation Workshop, vol. 1, pp. 141\u2013146 (2017)"},{"key":"2_CR21","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., Soricut, R.: ALBERT: a lite BERT for self-supervised learning of language representations. In: Proceedings of the 8th International Conference on Learning Representations, ICLR 2020. OpenReview.net, Addis Ababa, Ethiopia (2020)"},{"issue":"4","key":"2_CR22","doi-asserted-by":"publisher","first-page":"1234","DOI":"10.1093\/bioinformatics\/btz682","volume":"36","author":"J Lee","year":"2020","unstructured":"Lee, J., et al.: BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36(4), 1234\u20131240 (2020)","journal-title":"Bioinformatics"},{"key":"2_CR23","doi-asserted-by":"crossref","unstructured":"Liu, W., et al.: K-BERT: enabling language representation with knowledge graph. In: The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, pp. 2901\u20132908. AAAI Press, New York, NY, USA (2020)","DOI":"10.1609\/aaai.v34i03.5681"},{"key":"2_CR24","unstructured":"Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., et al.: RoBERTa: a robustly optimized BERT pretraining approach (2019)"},{"key":"2_CR25","unstructured":"Louis, A.: NetBERT: a pre-trained language representation model for computer networking, Master\u2019s thesis, University of Li\u00e8ge, Li\u00e8ge, Belgium (2020)"},{"key":"2_CR26","unstructured":"Lu, S., Guo, D., Ren, S., Huang, J., Svyatkovskiy, A., Blanco, A., et al.: CodeXGLUE: a machine learning benchmark dataset for code understanding and generation. arXiv preprint arXiv:2102.04664 (2021)"},{"key":"2_CR27","doi-asserted-by":"crossref","unstructured":"McCloskey, M., Cohen, N.J.: Catastrophic interference in connectionist networks: the sequential learning problem. In: Psychology of Learning and Motivation, vol. 24, pp. 109\u2013165 (1989)","DOI":"10.1016\/S0079-7421(08)60536-8"},{"key":"2_CR28","doi-asserted-by":"crossref","unstructured":"Nentidis, A., Bougiatiotis, K., Krithara, A., Paliouras, G.: Results of the seventh edition of the BioASQ challenge. In: Machine Learning and Knowledge Discovery in Databases: International Workshops of ECML PKDD 2019, W\u00fcrzburg, Germany, 16\u201320 September 2019, Proceedings, Part II, pp. 553\u2013568 (2020)","DOI":"10.1007\/978-3-030-43887-6_51"},{"key":"2_CR29","doi-asserted-by":"crossref","unstructured":"Nye, B.E., et al.: A corpus with multi-level annotations of patients, interventions and outcomes to support language processing for medical literature. In: Gurevych, I., Miyao, Y. (eds.) Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018, Melbourne, Australia, July 15\u201320, 2018, Volume 1: Long Papers, pp. 197\u2013207. Association for Computational Linguistics (2018)","DOI":"10.18653\/v1\/P18-1019"},{"issue":"10","key":"2_CR30","doi-asserted-by":"publisher","first-page":"1345","DOI":"10.1109\/TKDE.2009.191","volume":"22","author":"SJ Pan","year":"2010","unstructured":"Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22(10), 1345\u20131359 (2010)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"2_CR31","doi-asserted-by":"crossref","unstructured":"Peng, Y., Yan, S., Lu, Z.: Transfer learning in biomedical natural language processing: An evaluation of BERT and ELMO on ten benchmarking datasets. In: Proceedings of the 18th BioNLP Workshop and Shared Task, pp. 58\u201365. Association for Computational Linguistics, Florence, Italy (2019)","DOI":"10.18653\/v1\/W19-5006"},{"key":"2_CR32","volume-title":"Improving language understanding by generative pre-training","author":"A Radford","year":"2018","unstructured":"Radford, A., Narasimhan, K., Salimans, T., Sutskever, I., et al.: Improving language understanding by generative pre-training. Tech. rep, OpenAI (2018)"},{"issue":"8","key":"2_CR33","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)","journal-title":"OpenAI Blog"},{"key":"2_CR34","first-page":"1","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21, 1\u201367 (2020)","journal-title":"J. Mach. Learn. Res."},{"issue":"14","key":"2_CR35","doi-asserted-by":"publisher","first-page":"i49","DOI":"10.1093\/bioinformatics\/btx238","volume":"33","author":"G Sogancioglu","year":"2017","unstructured":"Sogancioglu, G., \u00d6zt\u00fcrk, H., \u00d6zg\u00fcr, A.: BIOSSES: a semantic sentence similarity estimation system for the biomedical domain. Bioinformatics 33(14), i49\u2013i58 (2017)","journal-title":"Bioinformatics"},{"key":"2_CR36","doi-asserted-by":"crossref","unstructured":"Tai, W., Kung, H.T., Dong, X., Comiter, M., Kuo, C.F.: exBERT: extending pre-trained models with domain-specific vocabulary under constrained training resources. In: Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 1433\u20131439. Association for Computational Linguistics (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.129"},{"key":"2_CR37","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., et al.: Attention is all you need. In: Guyon, I., et al. (eds.) Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, pp. 5998\u20136008, 4\u20139 December 2017, Long Beach, CA, USA (2017)"},{"key":"2_CR38","doi-asserted-by":"crossref","unstructured":"Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., Bowman, S.R.: GLUE: a multi-task benchmark and analysis platform for natural language understanding. In: 7th International Conference on Learning Representations, ICLR 2019. OpenReview.net, New Orleans, LA, USA (2019)","DOI":"10.18653\/v1\/W18-5446"},{"key":"2_CR39","unstructured":"Wu, Y., Schuster, M., Chen, Z., Le, Q.V., Norouzi, M., Macherey, W., et al.: Google\u2019s neural machine translation system: bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144 (2016)"},{"key":"2_CR40","unstructured":"Yang, Y., Uy, M.C.S., Huang, A.: FinBERT: a pretrained language model for financial communications. arXiv preprint arXiv:2006.08097 (2020)"},{"key":"2_CR41","doi-asserted-by":"crossref","unstructured":"Yao, Y., Huang, S., Wang, W., Dong, L., Wei, F.: Adapt-and-distill: developing small, fast and effective pretrained language models for domains. In: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pp. 460\u2013470. Association for Computational Linguistics (2021)","DOI":"10.18653\/v1\/2021.findings-acl.40"},{"key":"2_CR42","doi-asserted-by":"crossref","unstructured":"Yu, W., et al.: Dict-BERT: enhancing language model pre-training with dictionary. In: Findings of the Association for Computational Linguistics: ACL 2022, pp. 1907\u20131918. Association for Computational Linguistics, Dublin, Ireland (2022)","DOI":"10.18653\/v1\/2022.findings-acl.150"},{"key":"2_CR43","doi-asserted-by":"crossref","unstructured":"Zhu, H., Peng, H., Lyu, Z., Hou, L., Li, J., Xiao, J.: Pre-training language model incorporating domain-specific heterogeneous knowledge into a unified representation. Expert Systems with Applications, p. 119369 (2022)","DOI":"10.1016\/j.eswa.2022.119369"}],"container-title":["Lecture Notes in Computer Science","Knowledge Science, Engineering and Management"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-40292-0_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,25]],"date-time":"2024-10-25T22:09:24Z","timestamp":1729894164000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-40292-0_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031402913","9783031402920"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-40292-0_2","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"9 August 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"KSEM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Knowledge Science, Engineering and Management","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Guangzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 August 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 August 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ksem2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.ksem2023.conferences.academy\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"395","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"114","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"30","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"29% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2,5","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}