{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,9]],"date-time":"2026-03-09T01:16:04Z","timestamp":1773018964741,"version":"3.50.1"},"reference-count":27,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"4","license":[{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Pervasive Comput."],"published-print":{"date-parts":[[2024,10]]},"DOI":"10.1109\/mprv.2024.3499513","type":"journal-article","created":{"date-parts":[[2025,1,31]],"date-time":"2025-01-31T18:38:42Z","timestamp":1738348722000},"page":"87-93","source":"Crossref","is-referenced-by-count":7,"title":["Multimodal Large Language Models in Human-Centered Health: Practical Insights"],"prefix":"10.1109","volume":"23","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3806-1493","authenticated-orcid":false,"given":"Ting","family":"Dang","sequence":"first","affiliation":[{"name":"The University of Melbourne, Parkville, VIC, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6047-4158","authenticated-orcid":false,"given":"Hong","family":"Jia","sequence":"additional","affiliation":[{"name":"The University of Melbourne, Parkville, VIC, Australia"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1098\/rsos.230806"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-024-02855-5"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-023-00970-0"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1038\/d41586-023-00816-5"},{"key":"ref5","article-title":"LLaMA: Open and efficient foundation language models","volume-title":"Meta AI Res.","author":"Touvron","year":"2023"},{"key":"ref6","article-title":"ECG semantic integrator (ESI): A foundation ECG model pretrained with LLM-enhanced cardiological text","volume-title":"Proc. Trans. Mach. Learn. Res.","author":"Yu","year":"2024"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.133"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/3675094.3678494"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1145\/3675094.3678420"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1145\/3675094.3677547"},{"key":"ref11","article-title":"Health-LLM: Large language models for health prediction via wearable sensor data","volume-title":"Proc. Conf Health, Inference, Learn.","author":"Kim","year":"2024"},{"key":"ref12","article-title":"Visual instruction tuning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Liu","year":"2024"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.metrad.2024.100103"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-47679-2_7"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2024.3475471"},{"key":"ref16","article-title":"LoRA: Low-rank adaptation of large language models","author":"Hu","year":"2022","journal-title":"Proc. Int. Conf. Learn. Representations"},{"key":"ref17","article-title":"CALF: Aligning LLMs for time series forecasting via cross-modal fine-tuning","author":"Liu","year":"2024"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.418"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP48485.2024.10446343"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-naacl.231"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i17.29902"},{"key":"ref22","article-title":"Time-LLM: Time series forecasting by reprogramming large language models","author":"Jin","year":"2023","journal-title":"Proc. Int. Conf. Learn. Representations"},{"key":"ref23","article-title":"Listen, think, and understand","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Gong","year":"2024"},{"key":"ref24","article-title":"Exploring the reasoning abilities of multimodal large language models (MLLMs): A comprehensive survey on emerging trends in multimodal reasoning","author":"Wang","year":"2024"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-024-01083-y"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/s44336-024-00009-2"},{"key":"ref27","article-title":"Introducing gemini: Our largest and most capable ai model","year":"2024"}],"container-title":["IEEE Pervasive Computing"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/7756\/10859196\/10859301.pdf?arnumber=10859301","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,6]],"date-time":"2025-02-06T18:47:27Z","timestamp":1738867647000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10859301\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10]]},"references-count":27,"journal-issue":{"issue":"4"},"URL":"https:\/\/doi.org\/10.1109\/mprv.2024.3499513","relation":{},"ISSN":["1536-1268","1558-2590"],"issn-type":[{"value":"1536-1268","type":"print"},{"value":"1558-2590","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10]]}}}