{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T16:06:52Z","timestamp":1774541212017,"version":"3.50.1"},"reference-count":32,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"8","license":[{"start":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T00:00:00Z","timestamp":1722470400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T00:00:00Z","timestamp":1722470400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-009"},{"start":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T00:00:00Z","timestamp":1722470400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-001"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Computer"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1109\/mc.2024.3404674","type":"journal-article","created":{"date-parts":[[2024,8,9]],"date-time":"2024-08-09T17:21:23Z","timestamp":1723224083000},"page":"96-100","source":"Crossref","is-referenced-by-count":10,"title":["Compacting AI: In Search of the Small Language Model"],"prefix":"10.1109","volume":"57","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2355-9424","authenticated-orcid":false,"given":"Mla\u0111an","family":"Jovanovi\u0107","sequence":"first","affiliation":[{"name":"Singidunum University, Belgrade, Serbia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5415-6631","authenticated-orcid":false,"given":"Mark","family":"Campbell","sequence":"additional","affiliation":[{"name":"EVOTEK, San Diego, CA, USA"}]}],"member":"263","reference":[{"issue":"253","key":"ref1","first-page":"1","article-title":"Estimating carbon footprint of BLOOM, a 176B parameter language model","volume":"24","author":"Luccioni","year":"2023","journal-title":"J. Mach. Learn. Res."},{"key":"ref2","volume-title":"ShortGPT: Layers in large language models are more redundant than you expect","author":"Men","year":"2024"},{"key":"ref3","volume-title":"WeightWatcher: A diagnostic tool for deep neural networks","year":"2024"},{"key":"ref4","volume-title":"What tools are there to reduce AI power consumption?","author":"Shepard","year":"2024"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1038\/d41586-024-00478-x"},{"key":"ref6","volume-title":"Textbooks are all you need","author":"Gunasekar","year":"2023"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-acl.507"},{"key":"ref8","volume-title":"TinyLlama: An open-source small language model","author":"Zhang","year":"2024"},{"key":"ref9","volume-title":"Tools for building LLM applications","author":"Aishwarya","year":"2024"},{"key":"ref10","volume-title":"What I learned from looking at 900 most popular open source AI tools","author":"Huyen","year":"2024"},{"key":"ref11","volume-title":"GaLore: Memory-efficient LLM training by gradient low-rank projection","author":"Zhao","year":"2024"},{"key":"ref12","volume-title":"Llama 2: Open foundation and fine-tuned chat models","author":"Touvron","year":"2023"},{"key":"ref13","volume-title":"The falcon series of open language models","author":"Almazrouei","year":"2023"},{"key":"ref14","volume-title":"Mistral 7B","author":"Jiang","year":"2023"},{"key":"ref15","volume-title":"Gemma: Open models based on Gemini","year":"2024"},{"key":"ref16","volume-title":"Phi-3 technical report: A highly capable language model locally on your phone","author":"Abdin","year":"2024"},{"key":"ref17","volume-title":"OpenELM: An efficient language model family with open training and inference framework","author":"Mehta","year":"2024"},{"key":"ref18","volume-title":"Yi: Open foundation models by 01.AI","author":"Young","year":"2024"},{"key":"ref19","volume-title":"Aya model: An instruction finetuned open-access multilingual language model","author":"stn","year":"2024"},{"key":"ref20","volume-title":"OLMo: Accelerating the science of language models","author":"Groeneveld","year":"2024"},{"key":"ref21","volume-title":"7-step guide for small language models on local CPUs","author":"Sakshi","year":"2024"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/mcom.001.2300550"},{"key":"ref23","volume-title":"private communication","author":"Pham","year":"2024"},{"key":"ref24","volume-title":"Why larger language models do in-context learning differently?","author":"Shi","year":"2023"},{"key":"ref25","volume-title":"The era of 1-Bit LLMs: All large language models are in 1.58 bits","author":"Ma","year":"2024"},{"key":"ref26","volume-title":"Fine-tuning or retrieval? Comparing knowledge injection in LLMs","author":"Ovadia","year":"2024"},{"key":"ref27","volume-title":"Lost in the middle: How language models use long contexts","author":"Liu","year":"2023"},{"key":"ref28","volume-title":"Data authenticity, consent, & provenance for AI are all broken: What will it take to fix them?","author":"Longpre","year":"2024"},{"key":"ref29","volume-title":"Scalable extraction of training data from (production) language models","author":"Nasr","year":"2023"},{"key":"ref30","volume-title":"Orca 2: Teaching small language models how to reason","author":"Mitra","year":"2023"},{"key":"ref31","volume-title":"Talking about large language models","author":"Shanahan","year":"2023"},{"key":"ref32","volume-title":"Progress tracker: Open vs. proprietary LLMs","author":"Reed","year":"2024"}],"container-title":["Computer"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/2\/10632595\/10632613.pdf?arnumber=10632613","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,14]],"date-time":"2024-08-14T17:34:38Z","timestamp":1723656878000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10632613\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8]]},"references-count":32,"journal-issue":{"issue":"8"},"URL":"https:\/\/doi.org\/10.1109\/mc.2024.3404674","relation":{},"ISSN":["0018-9162","1558-0814"],"issn-type":[{"value":"0018-9162","type":"print"},{"value":"1558-0814","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,8]]}}}