{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T21:47:02Z","timestamp":1757627222832,"version":"3.44.0"},"publisher-location":"Cham","reference-count":12,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031948916"},{"type":"electronic","value":"9783031948923"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-94892-3_31","type":"book-chapter","created":{"date-parts":[[2025,8,17]],"date-time":"2025-08-17T07:16:01Z","timestamp":1755414961000},"page":"420-432","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Are Small Language Models Enough for\u00a0Biomedical QA Tasks?"],"prefix":"10.1007","author":[{"given":"Javier Lamar","family":"L\u00e9on","sequence":"first","affiliation":[]},{"given":"Vitor Beires","family":"Nogueira","sequence":"additional","affiliation":[]},{"given":"Paulo","family":"Quaresma","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,18]]},"reference":[{"key":"31_CR1","doi-asserted-by":"crossref","unstructured":"Aghajanyan, A., Gupta, S., Zettlemoyer, L.: Intrinsic dimensionality explains the effectiveness of language model fine-tuning. In: Zong, C., Xia, F., Li, W., Navigli, R. (eds.) Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 7319\u20137328. Association for Computational Linguistics (2021). https:\/\/aclanthology.org\/2021.acl-long.568\/","DOI":"10.18653\/v1\/2021.acl-long.568"},{"key":"31_CR2","unstructured":"Han, T., et al.: Medalpaca \u2013 an open-source collection of medical conversational AI models and training data (2023). https:\/\/arxiv.org\/abs\/2304.08247"},{"key":"31_CR3","unstructured":"Han, Z., Gao, C., Liu, J., Zhang, J., Zhang, S.Q.: Parameter-efficient fine-tuning for large models: a comprehensive survey (2024). https:\/\/arxiv.org\/abs\/2403.14608"},{"key":"31_CR4","doi-asserted-by":"crossref","unstructured":"Jin, Q., Dhingra, B., Liu, Z., Cohen, W., Lu, X.: PubMedQA: a dataset for biomedical research question answering. In: Inui, K., Jiang, J., Ng, V., Wan, X. (eds.) Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 2567\u20132577. Association for Computational Linguistics, Hong Kong (2019). https:\/\/aclanthology.org\/D19-1259\/","DOI":"10.18653\/v1\/D19-1259"},{"issue":"3","key":"31_CR5","first-page":"1","volume":"2","author":"H Li","year":"2024","unstructured":"Li, H., et al.: Codes: towards building open-source language models for text-to-SQL. Proc. ACM Manag. Data 2(3), 1\u201328 (2024)","journal-title":"Proc. ACM Manag. Data"},{"key":"31_CR6","unstructured":"Li, X.L., Liang, P.: Prefix-tuning: optimizing continuous prompts for generation (2021). https:\/\/arxiv.org\/abs\/2101.00190"},{"key":"31_CR7","unstructured":"Nori, H., et al.: Can generalist foundation models outcompete special-purpose tuning? case study in medicine (2023). https:\/\/arxiv.org\/abs\/2311.16452"},{"issue":"12","key":"31_CR8","doi-asserted-by":"publisher","first-page":"1418","DOI":"10.1038\/s42256-024-00944-1","volume":"6","author":"J Qiu","year":"2024","unstructured":"Qiu, J., et al.: LLM-based agentic systems in medicine and healthcare. Nature Mach. Intell. 6(12), 1418\u20131420 (2024)","journal-title":"Nature Mach. Intell."},{"key":"31_CR9","unstructured":"Sun, M.: Llamacare: A large medical language model for enhancing healthcare knowledge sharing (2024). https:\/\/arxiv.org\/abs\/2406.02350"},{"key":"31_CR10","unstructured":"Touvron, H., et al.: Llama 2: open foundation and fine-tuned chat models (2023). https:\/\/arxiv.org\/abs\/2307.09288"},{"key":"31_CR11","unstructured":"Wang, S., Yu, L., Li, J.: LoRA-GA: Low-rank adaptation with gradient approximation (2024). https:\/\/arxiv.org\/abs\/2407.05000"},{"key":"31_CR12","doi-asserted-by":"crossref","unstructured":"Yang, H., Li, S., Gon\u00e7alves, T.: Enhancing biomedical question answering with large language models. Information 15(8) (2024). https:\/\/www.mdpi.com\/2078-2489\/15\/8\/494","DOI":"10.3390\/info15080494"}],"container-title":["Lecture Notes in Networks and Systems","Proceedings of 17th International Conference on Machine Learning and Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-94892-3_31","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,10]],"date-time":"2025-09-10T01:26:39Z","timestamp":1757467599000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-94892-3_31"}},"subtitle":["Fine-Tuning Mistral-7B Using LoRA and PubMedQA"],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031948916","9783031948923"],"references-count":12,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-94892-3_31","relation":{},"ISSN":["2367-3370","2367-3389"],"issn-type":[{"type":"print","value":"2367-3370"},{"type":"electronic","value":"2367-3389"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"18 August 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICMLC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Machine Learning and Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Guangzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14 February 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 February 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icmlc-12025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.icmlc.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}