{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,29]],"date-time":"2024-10-29T11:11:22Z","timestamp":1730200282034,"version":"3.28.0"},"reference-count":31,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,12,5]],"date-time":"2023-12-05T00:00:00Z","timestamp":1701734400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,12,5]],"date-time":"2023-12-05T00:00:00Z","timestamp":1701734400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,12,5]]},"DOI":"10.1109\/bibm58861.2023.10385287","type":"proceedings-article","created":{"date-parts":[[2024,1,18]],"date-time":"2024-01-18T13:27:43Z","timestamp":1705584463000},"page":"2990-2995","source":"Crossref","is-referenced-by-count":1,"title":["Enhancing Medical Language Understanding: Adapting LLMs to the Medical Domain through Hybrid Granularity Mask Learning"],"prefix":"10.1109","author":[{"given":"Longjun","family":"Fan","sequence":"first","affiliation":[{"name":"Beijing University of Posts and Telecommunications,State Key Laboratory of Networking and Switching Technology,Beijing,China,100876"}]},{"given":"Xiaohong","family":"Liu","sequence":"additional","affiliation":[{"name":"Tsinghua University,Department of Computer Science and Technology,Beijing,China,100084"}]},{"given":"Yuhao","family":"Wang","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications,State Key Laboratory of Networking and Switching Technology,Beijing,China,100876"}]},{"given":"Guoxing","family":"Yang","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications,State Key Laboratory of Networking and Switching Technology,Beijing,China,100876"}]},{"given":"Zongxin","family":"Du","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications,State Key Laboratory of Networking and Switching Technology,Beijing,China,100876"}]},{"given":"Guangyu","family":"Wang","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications,State Key Laboratory of Networking and Switching Technology,Beijing,China,100876"}]}],"member":"263","reference":[{"key":"ref1","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"article-title":"Palm: Scaling language modeling with pathways","year":"2022","author":"Chowdhery","key":"ref2"},{"article-title":"Llama: Open and efficient foundation language models","year":"2023","author":"Touvron","key":"ref3"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-023-02552-9"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2023.104417"},{"article-title":"Beyond one-model-fits-all: A survey of domain specialization for large language models","year":"2023","author":"Ling","key":"ref6"},{"article-title":"Doctorglm: Fine-tuning your chinese doctor is not a herculean task","year":"2023","author":"Xiong","key":"ref7"},{"article-title":"Huatuo: Tuning llama model with chinese medical knowledge","year":"2023","author":"Wang","key":"ref8"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-023-00626-4"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.coling-main.603"},{"key":"ref11","first-page":"8","article-title":"Literature survey: domain adaptation algorithms for natural language processing","author":"Li","year":"2012","journal-title":"Department of Computer Science The Graduate Center, The City University of New York"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.26"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-short.149"},{"article-title":"Fine-tuning a llm using reinforcement learning from human feedback for a therapy chatbot application","year":"2023","author":"Bill","key":"ref14"},{"article-title":"Velma: Verbalization embodiment of llm agents for vision and language navigation in street view","year":"2023","author":"Schumann","key":"ref15"},{"article-title":"LoRA: Low-rank adaptation of large language models","volume-title":"International Conference on Learning Representations","author":"Hu","key":"ref16"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-short.8"},{"article-title":"Qlora: Efficient finetuning of quantized llms","year":"2023","author":"Dettmers","key":"ref18"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.754"},{"key":"ref20","first-page":"27730","article-title":"Training language models to follow instructions with human feedback","volume":"35","author":"Ouyang","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1017\/ATSIP.2019.12"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1139"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.740"},{"article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","year":"2018","author":"Devlin","key":"ref24"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.820"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.eacl-main.39"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1165"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.findings-emnlp.160"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.findings-acl.121"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2018.2883637"},{"article-title":"Glm-130b: An open bilingual pre-trained model","year":"2022","author":"Zeng","key":"ref31"}],"event":{"name":"2023 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)","start":{"date-parts":[[2023,12,5]]},"location":"Istanbul, Turkiye","end":{"date-parts":[[2023,12,8]]}},"container-title":["2023 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10385250\/10385251\/10385287.pdf?arnumber=10385287","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,19]],"date-time":"2024-01-19T13:27:10Z","timestamp":1705670830000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10385287\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,5]]},"references-count":31,"URL":"https:\/\/doi.org\/10.1109\/bibm58861.2023.10385287","relation":{},"subject":[],"published":{"date-parts":[[2023,12,5]]}}}