{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T18:36:44Z","timestamp":1775932604753,"version":"3.50.1"},"reference-count":40,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,12,19]],"date-time":"2024-12-19T00:00:00Z","timestamp":1734566400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2024,12,19]],"date-time":"2024-12-19T00:00:00Z","timestamp":1734566400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/100014718","name":"Innovative Research Group Project of the National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U22A20261"],"award-info":[{"award-number":["U22A20261"]}],"id":[{"id":"10.13039\/100014718","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100014718","name":"Innovative Research Group Project of the National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61402210"],"award-info":[{"award-number":["61402210"]}],"id":[{"id":"10.13039\/100014718","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62266037"],"award-info":[{"award-number":["62266037"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2025,1]]},"DOI":"10.1007\/s40747-024-01641-7","type":"journal-article","created":{"date-parts":[[2024,12,19]],"date-time":"2024-12-19T08:53:08Z","timestamp":1734598388000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["T-LLaMA: a Tibetan large language model based on LLaMA2"],"prefix":"10.1007","volume":"11","author":[{"given":"Hui","family":"Lv","sequence":"first","affiliation":[]},{"given":"Chi","family":"Pu","sequence":"additional","affiliation":[]},{"given":"La","family":"Duo","sequence":"additional","affiliation":[]},{"given":"Yan","family":"Li","sequence":"additional","affiliation":[]},{"given":"Qingguo","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Jun","family":"Shen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,19]]},"reference":[{"key":"1641_CR1","unstructured":"OpenAI. (2022) ChatGPT [GPT-3.5]. https:\/\/openai.com\/chatgpt"},{"key":"1641_CR2","unstructured":"Touvron H, Lavril T, Izacard G, Martinet X, Lachaux MA, Lacroix T, Lample G (2023) Llama: Open and efficient foundation language models. arXiv preprint. arXiv: 2302.13971"},{"key":"1641_CR3","unstructured":"Bai J, Bai S, Chu Y, Cui Z, Dang K, Deng X, Zhu T (2023) Qwen technical report. arXiv preprint. arXiv: 2309.16609"},{"key":"1641_CR4","unstructured":"Cui Y, Yang Z, Yao X (2023) Efficient and effective text encoding for chinese llama and alpaca. arXiv preprint. arXiv: 2304.08177"},{"key":"1641_CR5","unstructured":"Touvron H, Martin L, Stone K, Albert P, Almahairi A, Babaei Y, Scialom T (2023) Llama 2: Open foundation and fine-tuned chat models. arXiv preprint. arXiv: 2307.09288"},{"key":"1641_CR6","unstructured":"Hu EJ, Shen Y, Wallis P, Allen-Zhu Z, Li Y, Wang S, Chen W (2021) Lora: Low-rank adaptation of large language models. arXiv preprint. arXiv: 2106.09685"},{"key":"1641_CR7","doi-asserted-by":"crossref","unstructured":"Kudo T, Richardson J (2018) Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. arXiv preprint. arXiv: 1808.06226","DOI":"10.18653\/v1\/D18-2012"},{"key":"1641_CR8","doi-asserted-by":"crossref","unstructured":"Qun N, Li X, Qiu X, Huang X (2017) End-to-end neural text classification for tibetan. In: Chinese computational linguistics and natural language processing based on naturally annotated big data, pp 472\u2013480","DOI":"10.1007\/978-3-319-69005-6_39"},{"key":"1641_CR9","unstructured":"Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, Amodei D (2020) Language models are few-shot learners. In: Advances in neural information processing systems, pp 1877\u20131901"},{"issue":"240","key":"1641_CR10","first-page":"1","volume":"24","author":"A Chowdhery","year":"2023","unstructured":"Chowdhery A, Narang S, Devlin J, Bosma M, Mishra G, Roberts A, Fiedel N (2023) Palm: scaling language modeling with pathways. J Mach Learn Res 24(240):1\u2013113","journal-title":"J Mach Learn Res"},{"key":"1641_CR11","doi-asserted-by":"crossref","unstructured":"Sun X, Li X, Li J, Wu F, Guo S, Zhang T, Wang G (2023) Text classification via large language models.\u00a0arXiv preprint. arXiv: 2305.08377","DOI":"10.18653\/v1\/2023.findings-emnlp.603"},{"key":"1641_CR12","unstructured":"Kasai J, Sakaguchi K, Le Bras R, Asai A, Yu X, Radev D, Inui K (2024) RealTime QA: what's the answer right now?.\u00a0In: Advances in neural information processing systems, vol 36"},{"key":"1641_CR13","doi-asserted-by":"crossref","unstructured":"Kamalloo E, Dziri N, Clarke CL, Rafiei D (2023) Evaluating open-domain question answering in the era of large language models.\u00a0arXiv preprint. arXiv: 2305.06984","DOI":"10.18653\/v1\/2023.acl-long.307"},{"key":"1641_CR14","unstructured":"Zhao WX, Zhou K, Li J, Tang T, Wang X, Hou Y, Wen JR (2023) A survey of large language models. arXiv preprint. arXiv: 2303.18223"},{"key":"1641_CR15","doi-asserted-by":"crossref","unstructured":"Xiao L, Shan X, Chen X (2023) PatternGPT: a pattern-driven framework for large language model text generation. In: Proceedings of the 2023 12th International Conference on Computing and Pattern Recognition, pp 72\u201378","DOI":"10.1145\/3633637.3633648"},{"key":"1641_CR16","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2022.109460","volume":"252","author":"Q Xie","year":"2022","unstructured":"Xie Q, Bishop JA, Tiwari P, Ananiadou S (2022) Pre-trained language models with domain knowledge for biomedical extractive summarization. Knowl-Based Syst 252:109460","journal-title":"Knowl-Based Syst"},{"key":"1641_CR17","unstructured":"Houlsby N, Giurgiu A, Jastrzebski S, Morrone B, De Laroussilhe Q, Gesmundo A, Gelly S (2019) Parameter-efficient transfer learning for NLP. In: International Conference on machine learning, pp 2790\u20132799"},{"key":"1641_CR18","unstructured":"Li XL, Liang P (2021) Prefix-tuning: Optimizing continuous prompts for generation. arXiv preprint. arXiv: 2101.00190"},{"key":"1641_CR19","doi-asserted-by":"crossref","unstructured":"Lester B, Al-Rfou R, Constant N (2021) The power of scale for parameter-efficient prompt tuning. arXiv preprint. arXiv: 2104.08691","DOI":"10.18653\/v1\/2021.emnlp-main.243"},{"key":"1641_CR20","doi-asserted-by":"crossref","unstructured":"Hedderich MA, Lange L, Adel H, Str\u00f6tgen J, Klakow D (2020) A survey on recent approaches for natural language processing in low-resource scenarios.\u00a0arXiv preprint. arXiv: 2010.12309","DOI":"10.18653\/v1\/2021.naacl-main.201"},{"key":"1641_CR21","unstructured":"Nguyen XP, Zhang W, Li X, Aljunied M, Tan Q, Cheng L, Bing L (2023) SeaLLMs\u2014large language models for Southeast Asia.\u00a0arXiv preprint. arXiv: 2312.00738"},{"key":"1641_CR22","unstructured":"Yang A, Xiao B, Wang B, Zhang B, Bian C, Yin C, Wu Z (2023) Baichuan 2: open large-scale language models.\u00a0arXiv preprint. arXiv: 2309.10305"},{"key":"1641_CR23","unstructured":"Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint. arXiv: 1810.04805"},{"key":"1641_CR24","unstructured":"Lan Z, Chen M, Goodman S, Gimpel K, Sharma P, Soricut R (2019) Albert: a lite bert for self-supervised learning of language representations.\u00a0arXiv preprint arXiv: 1909.11942"},{"key":"1641_CR25","unstructured":"Radford A, Narasimhan K, Salimans T, Sutskever I (2018) Improving language understanding by generative pre-training"},{"issue":"8","key":"1641_CR26","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I (2019) Language models are unsupervised multitask learners. OpenAI blog 1(8):9","journal-title":"OpenAI blog"},{"issue":"4","key":"1641_CR27","doi-asserted-by":"publisher","first-page":"709","DOI":"10.26599\/TST.2021.9010060","volume":"27","author":"S Jiang","year":"2021","unstructured":"Jiang S, Fu S, Lin N, Fu Y (2021) Pretrained models and evaluation data for the Khmer language. Tsinghua Sci Technol 27(4):709\u2013718","journal-title":"Tsinghua Sci Technol"},{"key":"1641_CR28","doi-asserted-by":"crossref","unstructured":"Alghamdi A, Duan X, Jiang W, Wang Z, Wu Y, Xia Q, Ghaddar A (2023) Aramus: pushing the limits of data and model scale for Arabic natural language processing. arXiv preprint. arXiv: 2306.06800.","DOI":"10.18653\/v1\/2023.findings-acl.181"},{"key":"1641_CR29","volume-title":"Tibetan pre-trained model based on ALBERT and its application","author":"L Li","year":"2020","unstructured":"Li L (2020) Tibetan pre-trained model based on ALBERT and its application. Lanzhou University"},{"key":"1641_CR30","doi-asserted-by":"crossref","unstructured":"Liu S, Deng J, Sun Y, Zhao X (2022) Tibert: Tibetan pre-trained language model. In: 2022 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp 2956\u20132961","DOI":"10.1109\/SMC53654.2022.9945074"},{"key":"1641_CR31","unstructured":"Yang Z, Xu Z, Cui Y, Wang B, Lin M, Wu D, Chen Z (2022) CINO: A Chinese minority pre-trained language model. arXiv preprint. arXiv: 2202.13558."},{"issue":"8","key":"1641_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3603168","volume":"22","author":"B An","year":"2023","unstructured":"An B (2023) Prompt-based for low-resource Tibetan text classification. ACM Trans Asian Low-Resour Lang Inform Process 22(8):1\u201313","journal-title":"ACM Trans Asian Low-Resour Lang Inform Process"},{"key":"1641_CR33","doi-asserted-by":"crossref","unstructured":"Zhou M, Daiqing Z, Qun N, Nyima T (2023) PEFTT: Parameter-Efficient Fine-Tuning for low-resource Tibetan pre-trained language models. arXiv preprint. arXiv: 2309.12109.","DOI":"10.1007\/978-3-031-72350-6_28"},{"key":"1641_CR34","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Polosukhin I (2017) Attention is all you need. Advances in neural information processing systems"},{"key":"1641_CR35","unstructured":"Hoffmann J, Borgeaud S, Mensch A, Buchatskaya E, Cai T, Rutherford E, Sifre L (2022) Training compute-optimal large language models.\u00a0arXiv preprint. arXiv: 2203.15556."},{"key":"1641_CR36","unstructured":"Cooper K (2023) OpenAI GPT-3: Everything you need to know. [Online]. Available: https:\/\/www.springboard.com\/blog\/data-science\/machine-learning-gpt-3-open-ai\/. Accessed 17 May 2023"},{"key":"1641_CR37","doi-asserted-by":"crossref","unstructured":"Yi Z, Wu Q, Yu J, Tang Y, Liu X, Peng L, Ma J (2022) Tibetan Syllable Prediction with Pre-trained Cross-lingual Language Model. In: 2022 IEEE 5th International Conference on computer and communication engineering technology (CCET), pp 162\u2013166","DOI":"10.1109\/CCET55412.2022.9906389"},{"key":"1641_CR38","doi-asserted-by":"crossref","unstructured":"Qun N, Li X, Qiu X, Huang X (2017) End-to-end neural text classification for Tibetan. In:\u00a0Chinese Computational Linguistics and Natural Language Processing Based on Naturally Annotated Big Data: 16th China National Conference, CCL 2017, and 5th International Symposium, pp 472\u2013480.","DOI":"10.1007\/978-3-319-69005-6_39"},{"key":"1641_CR39","unstructured":"Gerganov, Georgi (2022) llama.cpp. GitHub. https:\/\/github.com\/ggerganov\/llama.cpp."},{"key":"1641_CR40","doi-asserted-by":"crossref","unstructured":"Lin CY, Hovy E (2003) Automatic evaluation of summaries using n-gram co-occurrence statistics. In: Proceedings of the 2003 Human Language Technology Conference of the North American chapter of the association for computational linguistics, pp 150\u2013157.","DOI":"10.3115\/1073445.1073465"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01641-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-024-01641-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01641-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,30]],"date-time":"2025-01-30T20:18:13Z","timestamp":1738268293000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-024-01641-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,19]]},"references-count":40,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,1]]}},"alternative-id":["1641"],"URL":"https:\/\/doi.org\/10.1007\/s40747-024-01641-7","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,19]]},"assertion":[{"value":"30 December 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 August 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 December 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Corresponding authors declare on behalf of all authors that there is no conflict of interest. We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"72"}}