{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,7]],"date-time":"2026-05-07T03:56:53Z","timestamp":1778126213759,"version":"3.51.4"},"reference-count":38,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,11,18]],"date-time":"2024-11-18T00:00:00Z","timestamp":1731888000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2024,11,18]],"date-time":"2024-11-18T00:00:00Z","timestamp":1731888000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/100006108","name":"U.S. Department of Health & Human Services | NIH | National Center for Advancing Translational Sciences","doi-asserted-by":"publisher","award":["UL1TR00441901"],"award-info":[{"award-number":["UL1TR00441901"]}],"id":[{"id":"10.13039\/100006108","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["npj Digit. Med."],"DOI":"10.1038\/s41746-024-01315-1","type":"journal-article","created":{"date-parts":[[2024,11,18]],"date-time":"2024-11-18T10:01:50Z","timestamp":1731924110000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":33,"title":["A strategy for cost-effective large language model use at health system-scale"],"prefix":"10.1038","volume":"7","author":[{"given":"Eyal","family":"Klang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6217-1620","authenticated-orcid":false,"given":"Donald","family":"Apakama","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3295-7996","authenticated-orcid":false,"given":"Ethan E.","family":"Abbott","sequence":"additional","affiliation":[]},{"given":"Akhil","family":"Vaid","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0081-0356","authenticated-orcid":false,"given":"Joshua","family":"Lampert","sequence":"additional","affiliation":[]},{"given":"Ankit","family":"Sakhuja","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4946-6533","authenticated-orcid":false,"given":"Robert","family":"Freeman","sequence":"additional","affiliation":[]},{"given":"Alexander W.","family":"Charney","sequence":"additional","affiliation":[]},{"given":"David","family":"Reich","sequence":"additional","affiliation":[]},{"given":"Monica","family":"Kraft","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6319-4314","authenticated-orcid":false,"given":"Girish N.","family":"Nadkarni","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4515-8090","authenticated-orcid":false,"given":"Benjamin S.","family":"Glicksberg","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,18]]},"reference":[{"key":"1315_CR1","doi-asserted-by":"publisher","first-page":"bbad493","DOI":"10.1093\/bib\/bbad493","volume":"25","author":"S Tian","year":"2023","unstructured":"Tian, S. et al. Opportunities and challenges for ChatGPT and large language models in biomedicine and health. Brief. Bioinform. 25, bbad493 (2023).","journal-title":"Brief. Bioinform."},{"key":"1315_CR2","doi-asserted-by":"publisher","first-page":"1930","DOI":"10.1038\/s41591-023-02448-8","volume":"29","author":"AJ Thirunavukarasu","year":"2023","unstructured":"Thirunavukarasu, A. J. et al. Large language models in medicine. Nat. Med. 29, 1930\u20131940 (2023).","journal-title":"Nat. Med."},{"key":"1315_CR3","doi-asserted-by":"publisher","first-page":"194","DOI":"10.1038\/s41746-022-00742-2","volume":"5","author":"X Yang","year":"2022","unstructured":"Yang, X. et al. A large language model for electronic health records. Npj Digit. Med. 5, 194 (2022).","journal-title":"Npj Digit. Med."},{"key":"1315_CR4","doi-asserted-by":"publisher","first-page":"6","DOI":"10.1038\/s41746-023-00970-0","volume":"7","author":"M Guevara","year":"2024","unstructured":"Guevara, M. et al. Large language models to identify social determinants of health in electronic health records. Npj Digit. Med. 7, 6 (2024).","journal-title":"Npj Digit. Med."},{"key":"1315_CR5","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocae146","author":"M Sushil","year":"2024","unstructured":"Sushil, M. et al. A comparative study of large language model-based zero-shot inference and task-specific supervised classification of breast cancer pathology reports. J. Am. Med. Inform. Assoc. https:\/\/doi.org\/10.1093\/jamia\/ocae146 (2024).","journal-title":"J. Am. Med. Inform. Assoc."},{"key":"1315_CR6","doi-asserted-by":"publisher","first-page":"357","DOI":"10.1038\/s41586-023-06160-y","volume":"619","author":"LY Jiang","year":"2023","unstructured":"Jiang, L. Y. et al. Health system-scale language models are all-purpose prediction engines. Nature 619, 357\u2013362 (2023).","journal-title":"Nature"},{"key":"1315_CR7","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-024-45879-8","volume":"15","author":"S Tayebi Arasteh","year":"2024","unstructured":"Tayebi Arasteh, S. et al. Large language models streamline automated machine learning for clinical studies. Nat. Commun. 15, 1603 (2024).","journal-title":"Nat. Commun."},{"key":"1315_CR8","doi-asserted-by":"publisher","first-page":"e248895","DOI":"10.1001\/jamanetworkopen.2024.8895","volume":"7","author":"CYK Williams","year":"2024","unstructured":"Williams, C. Y. K. et al. Use of a large language model to assess clinical acuity of adults in the emergency department. JAMA Netw. Open 7, e248895 (2024).","journal-title":"JAMA Netw. Open"},{"key":"1315_CR9","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocae103","author":"BS Glicksberg","year":"2024","unstructured":"Glicksberg, B. S. et al. Evaluating the accuracy of a state-of-the-art large language model for prediction of admissions from the emergency room. J. Am. Med. Inform. Assoc. https:\/\/doi.org\/10.1093\/jamia\/ocae103 (2024).","journal-title":"J. Am. Med. Inform. Assoc."},{"key":"1315_CR10","doi-asserted-by":"publisher","first-page":"172","DOI":"10.1038\/s41586-023-06291-2","volume":"620","author":"K Singhal","year":"2023","unstructured":"Singhal, K. et al. Large language models encode clinical knowledge. Nature 620, 172\u2013180 (2023).","journal-title":"Nature"},{"key":"1315_CR11","doi-asserted-by":"publisher","first-page":"1134","DOI":"10.1038\/s41591-024-02855-5","volume":"30","author":"D Van Veen","year":"2024","unstructured":"Van Veen, D. et al. Adapted large language models can outperform medical experts in clinical text summarization. Nat. Med. 30, 1134\u20131142 (2024).","journal-title":"Nat. Med."},{"key":"1315_CR12","doi-asserted-by":"publisher","first-page":"e240357","DOI":"10.1001\/jamanetworkopen.2024.0357","volume":"7","author":"J Zaretsky","year":"2024","unstructured":"Zaretsky, J. et al. Generative artificial intelligence to transform inpatient discharge summaries to patient-friendly language and format. JAMA Netw. Open 7, e240357 (2024).","journal-title":"JAMA Netw. Open"},{"key":"1315_CR13","doi-asserted-by":"publisher","first-page":"e379","DOI":"10.1016\/S2589-7500(24)00060-8","volume":"6","author":"S Chen","year":"2024","unstructured":"Chen, S. et al. The effect of using a large language model to respond to patient messages. Lancet Digit. Health 6, e379\u2013e381 (2024).","journal-title":"Lancet Digit. Health"},{"key":"1315_CR14","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocad258","author":"S Tripathi","year":"2024","unstructured":"Tripathi, S., Sukumaran, R. & Cook, T. S. Efficient healthcare with large language models: optimizing clinical workflow and enhancing patient care. J. Am. Med. Inform. Assoc. https:\/\/doi.org\/10.1093\/jamia\/ocad258 (2024).","journal-title":"J. Am. Med. Inform. Assoc."},{"key":"1315_CR15","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocae073","author":"M Nievas","year":"2024","unstructured":"Nievas, M., Basu, A., Wang, Y. & Singh, H. Distilling large language models for matching patients to clinical trials. J. Am. Med. Inform. Assoc. https:\/\/doi.org\/10.1093\/jamia\/ocae073 (2024).","journal-title":"J. Am. Med. Inform. Assoc."},{"key":"1315_CR16","doi-asserted-by":"publisher","first-page":"120","DOI":"10.1038\/s41746-023-00873-0","volume":"6","author":"B Mesk\u00f3","year":"2023","unstructured":"Mesk\u00f3, B. & Topol, E. J. The imperative for regulatory oversight of large language models (or generative AI) in healthcare. Npj Digit. Med. 6, 120 (2023).","journal-title":"Npj Digit. Med."},{"key":"1315_CR17","doi-asserted-by":"publisher","first-page":"e2233348","DOI":"10.1001\/jamanetworkopen.2022.33348","volume":"5","author":"J Steinkamp","year":"2022","unstructured":"Steinkamp, J., Kantrowitz, J. J. & Airan-Javia, S. Prevalence and sources of duplicate information in the electronic medical record. JAMA Netw. Open 5, e2233348 (2022).","journal-title":"JAMA Netw. Open"},{"key":"1315_CR18","unstructured":"Sahoo, P. et al. A systematic survey of prompt engineering in large language models: techniques and applications. Preprint at arXiv http:\/\/arxiv.org\/abs\/2402.07927 (2024)."},{"key":"1315_CR19","unstructured":"White, J. et al. A prompt pattern catalog to enhance prompt engineering with ChatGPT. Preprint at arXiv http:\/\/arxiv.org\/abs\/2302.11382 (2023)."},{"key":"1315_CR20","unstructured":"Chang, K. et al. Efficient prompting methods for large language models: a survey. Preprint at arXiv http:\/\/arxiv.org\/abs\/2404.01077 (2024)."},{"key":"1315_CR21","unstructured":"Khot, T. et al. Decomposed prompting: a modular approach for solving complex tasks. Preprint at arXiv http:\/\/arxiv.org\/abs\/2210.02406 (2023)."},{"key":"1315_CR22","unstructured":"Wei, J. et al. Chain-of-thought prompting elicits reasoning in large language models. Preprint at arXiv http:\/\/arxiv.org\/abs\/2201.11903 (2023)."},{"key":"1315_CR23","unstructured":"Yugeswardeenoo, D., Zhu, K. & O\u2019Brien, S. Question-analysis prompting improves LLM performance in reasoning tasks. Preprint at arXiv http:\/\/arxiv.org\/abs\/2407.03624 (2024)."},{"key":"1315_CR24","unstructured":"Liu, N. F. et al. Lost in the middle: how language models use long contexts. Preprint at arXiv http:\/\/arxiv.org\/abs\/2307.03172 (2023)."},{"key":"1315_CR25","unstructured":"Li, T., Zhang, G., Do, Q. D., Yue, X. & Chen, W. Long-context LLMs struggle with long in-context learning. Preprint at http:\/\/arxiv.org\/abs\/2404.02060 (2024)."},{"key":"1315_CR26","unstructured":"Kuratov, Y. et al. In search of needles in a 11M Haystack: recurrent memory finds what LLMs miss. Preprint at arXiv http:\/\/arxiv.org\/abs\/2402.10790 (2024)."},{"key":"1315_CR27","doi-asserted-by":"crossref","unstructured":"Levy, M., Jacoby, A. & Goldberg, Y. Same task, more tokens: the impact of input length on the reasoning performance of large language models. Preprint at arXiv http:\/\/arxiv.org\/abs\/2402.14848 (2024).","DOI":"10.18653\/v1\/2024.acl-long.818"},{"key":"1315_CR28","doi-asserted-by":"publisher","first-page":"195","DOI":"10.1038\/s41746-023-00939-z","volume":"6","author":"JA Omiye","year":"2023","unstructured":"Omiye, J. A., Lester, J. C., Spichak, S., Rotemberg, V. & Daneshjou, R. Large language models propagate race-based medicine. NPJ Digit. Med. 6, 195 (2023).","journal-title":"NPJ Digit. Med."},{"key":"1315_CR29","doi-asserted-by":"crossref","unstructured":"Soroush, A. et al. Large language models are poor medical coders\u2014benchmarking of medical code querying. NEJM AI 1, AIdbp2300040 (2024).","DOI":"10.1056\/AIdbp2300040"},{"key":"1315_CR30","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1038\/s41746-023-00896-7","volume":"6","author":"L Tang","year":"2023","unstructured":"Tang, L. et al. Evaluating large language models on medical evidence summarization. NPJ Digit. Med. 6, 158 (2023).","journal-title":"NPJ Digit. Med."},{"key":"1315_CR31","unstructured":"Pal, A., Umapathi, L. K. & Sankarasubbu, M. MedMCQA: a large-scale multi-subject multi-choice dataset for medical domain question answering. in Proceedings of the Conference on Health, Inference, and Learning (eds Flores, G., Chen, G. H., Pollard, T., Ho, J. C. & Naumann, T.) 174 248\u2013260 (PMLR, 2022)."},{"key":"1315_CR32","unstructured":"OpenAI et al. GPT-4 Technical Report. Preprint at arXiv http:\/\/arxiv.org\/abs\/2303.08774 (2024)."},{"key":"1315_CR33","unstructured":"OpenAI. GPT3.5-Turbo. OpenAI."},{"key":"1315_CR34","unstructured":"AI@Meta. Llama 3 Model Card. (2024)."},{"key":"1315_CR35","unstructured":"Ankit Pal, M. S. OpenBioLLMs: advancing open-source large language models for healthcare and life sciences. Hugging Face Repository. https:\/\/huggingface.co\/blog\/aaditya\/openbiollm (2024)."},{"key":"1315_CR36","unstructured":"Jiang, A. Q. et al. Mixtral of Experts. Preprint at arXiv http:\/\/arxiv.org\/abs\/2401.04088 (2024)."},{"key":"1315_CR37","doi-asserted-by":"crossref","unstructured":"Labrak, Y. et al. BioMistral: a collection of open-source pretrained large language models for medical domains. Preprint at arXiv http:\/\/arxiv.org\/abs\/2402.10373 (2024).","DOI":"10.18653\/v1\/2024.findings-acl.348"},{"key":"1315_CR38","unstructured":"Gemma Team et al. Gemma: open models based on gemini research and technology. Preprint at arXiv http:\/\/arxiv.org\/abs\/2403.08295 (2024)."}],"container-title":["npj Digital Medicine"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s41746-024-01315-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-024-01315-1","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-024-01315-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,18]],"date-time":"2024-11-18T11:06:27Z","timestamp":1731927987000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s41746-024-01315-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,18]]},"references-count":38,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2024,12]]}},"alternative-id":["1315"],"URL":"https:\/\/doi.org\/10.1038\/s41746-024-01315-1","relation":{},"ISSN":["2398-6352"],"issn-type":[{"value":"2398-6352","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,18]]},"assertion":[{"value":"12 May 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 October 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 November 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Girish Nadkarni is an Associate Editor for NPJ Digital Medicine. The authors declare no competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"320"}}