{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,26]],"date-time":"2025-11-26T23:07:52Z","timestamp":1764198472519,"version":"3.46.0"},"publisher-location":"Singapore","reference-count":21,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819549597"},{"type":"electronic","value":"9789819549603"}],"license":[{"start":{"date-parts":[[2025,11,27]],"date-time":"2025-11-27T00:00:00Z","timestamp":1764201600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,27]],"date-time":"2025-11-27T00:00:00Z","timestamp":1764201600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-4960-3_2","type":"book-chapter","created":{"date-parts":[[2025,11,26]],"date-time":"2025-11-26T04:58:06Z","timestamp":1764133086000},"page":"14-26","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A Small Language Model and\u00a0Domain-Specific Resources for\u00a0Vietnamese Public Services"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-3917-6967","authenticated-orcid":false,"given":"Van Thai","family":"Le","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2645-1419","authenticated-orcid":false,"given":"Anh-Cuong","family":"Le","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,27]]},"reference":[{"key":"2_CR1","doi-asserted-by":"crossref","unstructured":"Aralimatti, R., Shakhadri, S.A.G., KR, K., Angadi, K.B.: Fine-tuning small language models for domain-specific AI: an edge AI perspective (2025). https:\/\/arxiv.org\/abs\/2503.01933","DOI":"10.20944\/preprints202502.2128.v1"},{"key":"2_CR2","doi-asserted-by":"crossref","unstructured":"Ariai, F., Demartini, G.: Natural language processing for the legal domain: a survey of tasks, datasets, models, and challenges (2025). https:\/\/arxiv.org\/abs\/2410.21306","DOI":"10.1145\/3777009"},{"key":"2_CR3","unstructured":"Dettmers, T., Lewis, M., Belkada, Y., Zettlemoyer, L.: LLM.int8(): 8-bit matrix multiplication for transformers at scale (2022). https:\/\/arxiv.org\/abs\/2208.07339"},{"key":"2_CR4","unstructured":"Gholami, S., Omar, M.: Do generative large language models need billions of parameters? ArXiv abs\/2309.06589 (2023). https:\/\/api.semanticscholar.org\/CorpusID:261705636"},{"key":"2_CR5","doi-asserted-by":"crossref","unstructured":"Gururangan, S., et al.: Don\u2019t stop pretraining: adapt language models to domains and tasks (2020). https:\/\/arxiv.org\/abs\/2004.10964","DOI":"10.18653\/v1\/2020.acl-main.740"},{"key":"2_CR6","unstructured":"Han, Z., Gao, C., Liu, J., Zhang, J., Zhang, S.Q.: Parameter-efficient fine-tuning for large models: a comprehensive survey (2024). https:\/\/arxiv.org\/abs\/2403.14608"},{"key":"2_CR7","unstructured":"Hu, E.J., et al.: Lora: low-rank adaptation of large language models (2021). https:\/\/arxiv.org\/abs\/2106.09685"},{"key":"2_CR8","unstructured":"Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Proceedings of the ACL Workshop on Text Summarization Branches Out, pp. 74\u201381 (2004)"},{"key":"2_CR9","unstructured":"Madaan, A., et al.: Self-refine: iterative refinement with self-feedback (2023). https:\/\/arxiv.org\/abs\/2303.17651"},{"key":"2_CR10","unstructured":"Minaee, S., et al.: Large language models: a survey (2025). https:\/\/arxiv.org\/abs\/2402.06196"},{"key":"2_CR11","doi-asserted-by":"crossref","unstructured":"Nguyen, D.Q., Nguyen, A.T.: PhoBERT: pre-trained language models for Vietnamese (2020). https:\/\/arxiv.org\/abs\/2003.00744","DOI":"10.18653\/v1\/2020.findings-emnlp.92"},{"issue":"2","key":"2_CR12","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1017\/nlp.2024.33","volume":"31","author":"P Pakray","year":"2025","unstructured":"Pakray, P., Gelbukh, A., Bandyopadhyay, S.: Natural language processing applications for low-resource languages. Nat. Lang. Process. 31(2), 183\u2013197 (2025). https:\/\/doi.org\/10.1017\/nlp.2024.33","journal-title":"Nat. Lang. Process."},{"key":"2_CR13","unstructured":"Pandita, D., Weerasooriya, T.C., Shah, A.P., Homan, C.M., Wei, W.: ProRefine: inference-time prompt refinement with textual feedback (2025). https:\/\/arxiv.org\/abs\/2506.05305"},{"key":"2_CR14","doi-asserted-by":"crossref","unstructured":"Phan, L., Tran, H., Nguyen, H., Trinh, T.H.: ViT5: pretrained text-to-text transformer for Vietnamese language generation (2022). https:\/\/arxiv.org\/abs\/2205.06457","DOI":"10.18653\/v1\/2022.naacl-srw.18"},{"key":"2_CR15","doi-asserted-by":"publisher","unstructured":"Ranaldi, L., Freitas, A.: Self-refine instruction-tuning for aligning reasoning in language models. In: Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 2325\u20132347. Association for Computational Linguistics (2024). https:\/\/doi.org\/10.18653\/v1\/2024.emnlp-main.139","DOI":"10.18653\/v1\/2024.emnlp-main.139"},{"key":"2_CR16","unstructured":"Tinn, R., et al.: Fine-tuning large neural language models for biomedical natural language processing (2021). https:\/\/arxiv.org\/abs\/2112.07869"},{"key":"2_CR17","series-title":"Studies in Computational Intelligence","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1007\/978-3-030-67008-5_2","volume-title":"Software Engineering, Artificial Intelligence, Networking and Parallel\/Distributed Computing","author":"T-T-T Vo","year":"2021","unstructured":"Vo, T.-T.-T., Van, H.-T.: Understanding factors influencing intention to use e-government services in Vietnam: focused on privacy and security concerns. In: Lee, R., Kim, J.B. (eds.) SNPD 2021. SCI, vol. 951, pp. 13\u201323. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-67008-5_2"},{"key":"2_CR18","unstructured":"Wan, Z., et al.: Efficient large language models: a survey (2024). https:\/\/arxiv.org\/abs\/2312.03863"},{"key":"2_CR19","unstructured":"Xu, H., Gan, W., Qi, Z., Wu, J., Yu, P.S.: Large language models for education: a survey (2024). https:\/\/arxiv.org\/abs\/2405.13001"},{"key":"2_CR20","unstructured":"Yang, A., et al.: Qwen3 technical report (2025). https:\/\/arxiv.org\/abs\/2505.09388"},{"key":"2_CR21","unstructured":"Zhang, T., Kishore, V., Wu, F., Weinberger, K.Q., Artzi, Y.: BERTScore: evaluating text generation with BERT (2020). https:\/\/arxiv.org\/abs\/1904.09675"}],"container-title":["Lecture Notes in Computer Science","Multi-disciplinary Trends in Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-4960-3_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,26]],"date-time":"2025-11-26T23:03:33Z","timestamp":1764198213000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-4960-3_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,27]]},"ISBN":["9789819549597","9789819549603"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-4960-3_2","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025,11,27]]},"assertion":[{"value":"27 November 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MIWAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multi-disciplinary Trends in Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Ho Chi Minh City","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vietnam","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 December 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 December 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miwai2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/miwai25.miwai.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}