{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,1,11]],"date-time":"2025-01-11T05:36:12Z","timestamp":1736573772164,"version":"3.32.0"},"reference-count":50,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"name":"Institute of Information communications Technology Planning Evaluation","award":["RS-2022-00155966"],"award-info":[{"award-number":["RS-2022-00155966"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/access.2024.3525041","type":"journal-article","created":{"date-parts":[[2025,1,2]],"date-time":"2025-01-02T19:32:30Z","timestamp":1735846350000},"page":"4957-4965","source":"Crossref","is-referenced-by-count":0,"title":["PRISM-Med: Parameter-Efficient Robust Interdomain Specialty Model for Medical Language Tasks"],"prefix":"10.1109","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0009-0000-7691-0930","authenticated-orcid":false,"given":"Jieui","family":"Kang","sequence":"first","affiliation":[{"name":"Artificial Intelligence Convergence, Ewha Womans University, Seoul, South Korea"}]},{"given":"Hyungon","family":"Ryu","sequence":"additional","affiliation":[{"name":"Nvidia Coporation, Seoul, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8722-8486","authenticated-orcid":false,"given":"Jaehyeong","family":"Sim","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Ewha Womans University, Seoul, South Korea"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Knowledge solver: Teaching LLMs to search for domain knowledge from knowledge graphs","author":"Feng","year":"2023","journal-title":"arXiv:2309.03118"},{"key":"ref2","article-title":"LoRA: Low-rank adaptation of large language models","author":"Hu","year":"2021","journal-title":"arXiv:2106.09685"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.02270"},{"key":"ref4","article-title":"Mixture of LoRa experts","author":"Wu","year":"2024","journal-title":"arXiv:2404.13628"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1016\/j.compind.2022.103612"},{"key":"ref6","article-title":"Flex-MoE: Modeling arbitrary modality combination via the flexible mixture-of-experts","author":"Yun","year":"2024","journal-title":"arXiv:2410.08245"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2019\/681"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102623"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D15-1049"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICMLA.2019.00155"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.173"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.740"},{"key":"ref13","article-title":"LoRa: Low-rank adaptation of large language models","author":"Hu","year":"2021","journal-title":"arXiv:2106.09685"},{"key":"ref14","first-page":"1","article-title":"A domain-adaptive pre-training approach for language bias detection in news","volume-title":"Proc. ACM\/IEEE Joint Conf. Digit. Libraries (JCDL)","author":"Krieger"},{"key":"ref15","article-title":"FDAPT: Federated domain-adaptive pre-training for language models","author":"Jiang","year":"2023","journal-title":"arXiv:2307.06933"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3131386"},{"key":"ref17","first-page":"65","article-title":"Automatically adapting an NLP core engine to the biology domain","volume-title":"Proc. ISMB","author":"Buyko"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-023-02448-8"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1145\/3616855.3635739"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-023-02412-6"},{"key":"ref21","first-page":"648","article-title":"Cross-lingual and cross-domain transfer learning for automatic term extraction from low resource data","volume-title":"Proc. 13th Lang. Resour. Eval. Conf.","author":"Hazem"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2015-700"},{"key":"ref23","article-title":"Chain of LoRA: Efficient fine-tuning of language models via residual learning","author":"Xia","year":"2024","journal-title":"arXiv:2401.04151"},{"key":"ref24","first-page":"10760","article-title":"LoNAS: Elastic low-rank adapters for efficient large language models","volume-title":"Proc. Joint Int. Conf. Comput. Linguistics, Lang. Resour. Eval.","author":"Munoz"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.3115\/1567594.1567618"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/BIBM58861.2023.10385678"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1186\/s12961-023-00972-0"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.126"},{"key":"ref29","article-title":"Think while you write: Hypothesis verification promotes faithful knowledge-to-text generation","author":"Qiu","year":"2023","journal-title":"arXiv:2311.09467"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/3410352.3410744"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i11.26596"},{"issue":"6","key":"ref32","first-page":"90","article-title":"Review on determining number of cluster in K-means clustering","volume":"1","author":"Kodinariya","year":"2013","journal-title":"Int. J."},{"issue":"140","key":"ref33","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2019","journal-title":"J. Mach. Learn. Res."},{"key":"ref34","article-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2014","journal-title":"arXiv:1412.6980"},{"key":"ref35","article-title":"Adapting large language models to domains via reading comprehension","author":"Cheng","year":"2023","journal-title":"arXiv:2309.09530"},{"volume-title":"Argilla\u2014Open-Source Framework for Data-Centric NLP","year":"2023","key":"ref36"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1186\/s12859-019-3119-4"},{"key":"ref38","first-page":"248","article-title":"MedMCQA: A large-scale multi-subject multi-choice dataset for medical domain question answering","volume-title":"Proc. Conf. Health, Inference, Learn.","author":"Pal"},{"volume-title":"PEFT: State-of-the-Art Parameter-Efficient Fine-Tuning Methods","year":"2022","author":"Mangrulkar","key":"ref39"},{"key":"ref40","article-title":"Continual learning in the frequency domain","author":"Liu","year":"2024","journal-title":"arXiv:2410.06645"},{"key":"ref41","article-title":"TinyLlama: An open-source small language model","author":"Zhang","year":"2024","journal-title":"arXiv:2401.02385"},{"issue":"3","key":"ref42","first-page":"3","article-title":"Phi-2: The surprising power of small language models","volume":"1","author":"Javaheripi","year":"2023","journal-title":"Microsoft Res. Blog"},{"key":"ref43","article-title":"Gemma: Open models based on Gemini research and technology","author":"Team","year":"2024","journal-title":"arXiv:2403.08295"},{"volume-title":"TRL: Transformer Reinforcement Learning","year":"2023","key":"ref44"},{"key":"ref45","article-title":"MEDITRON-70B: Scaling medical pretraining for large language models","author":"Chen","year":"2023","journal-title":"arXiv:2311.16079"},{"volume-title":"Med Llama2","year":"2023","author":"Raval","key":"ref46"},{"volume-title":"OpenBioLLMs: Advancing Open-Source Large Language Models for Healthcare and Life Sciences","year":"2024","author":"Ankit","key":"ref47"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-acl.348"},{"key":"ref49","article-title":"MedAlpaca\u2014An open-source collection of medical conversational AI models and training data","author":"Han","year":"2023","journal-title":"arXiv:2304.08247"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocae045"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10820123\/10820505.pdf?arnumber=10820505","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,10]],"date-time":"2025-01-10T20:50:16Z","timestamp":1736542216000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10820505\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":50,"URL":"https:\/\/doi.org\/10.1109\/access.2024.3525041","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2025]]}}}