{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,23]],"date-time":"2025-08-23T05:27:06Z","timestamp":1755926826322},"reference-count":30,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,10,1]],"date-time":"2023-10-01T00:00:00Z","timestamp":1696118400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,10,1]],"date-time":"2023-10-01T00:00:00Z","timestamp":1696118400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,10,1]]},"DOI":"10.1109\/smc53992.2023.10393961","type":"proceedings-article","created":{"date-parts":[[2024,1,29]],"date-time":"2024-01-29T18:32:04Z","timestamp":1706553124000},"source":"Crossref","is-referenced-by-count":3,"title":["MiLMo:Minority Multilingual Pre-Trained Language Model"],"prefix":"10.1109","author":[{"given":"Junjie","family":"Deng","sequence":"first","affiliation":[{"name":"School of information engineering, Minzu University of China,Beijing,100081"}]},{"given":"Hanru","family":"Shi","sequence":"additional","affiliation":[{"name":"School of information engineering, Minzu University of China,Beijing,100081"}]},{"given":"Xinhe","family":"Yu","sequence":"additional","affiliation":[{"name":"School of information engineering, Minzu University of China,Beijing,100081"}]},{"given":"Wugedele","family":"Bao","sequence":"additional","affiliation":[{"name":"Hohhot Minzu College,Hohhot,010000"}]},{"given":"Yuan","family":"Sun","sequence":"additional","affiliation":[{"name":"School of information engineering, Minzu University of China,Beijing,100081"}]},{"given":"Xiaobing","family":"Zhao","sequence":"additional","affiliation":[{"name":"School of information engineering, Minzu University of China,Beijing,100081"}]}],"member":"263","reference":[{"key":"ref1","first-page":"4171","article-title":"Bert: Pre-training of deep bidi-rectional transformers for language understanding","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019","volume":"1","author":"Devlin"},{"issue":"1","key":"ref2","first-page":"5485","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"The Journal of Machine Learning Research"},{"key":"ref3","author":"Liu","year":"2019","journal-title":"Roberta: A robustly optimized bert pretraining approach"},{"key":"ref4","author":"Radford","year":"2018","journal-title":"Improving language understanding by generative pre-training"},{"key":"ref5","first-page":"32","article-title":"Xlnet: Generalized autoregressive pre-training for language understanding","author":"Yang","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref6","first-page":"30","article-title":"Attention is all you need","author":"Vaswani","year":"2017","journal-title":"Advances in neural information processing systems"},{"issue":"8","key":"ref7","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI blog"},{"key":"ref8","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref9","article-title":"Albert: A lite bert for self-supervised learning of language representations","volume-title":"International Conference on Learning Representations","author":"Lan","year":"2019"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00300"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1493"},{"key":"ref12","first-page":"32","article-title":"Cross-lingual language model pretraining","author":"Conneau","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.747"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.3"},{"key":"ref15","first-page":"133","article-title":"KoNLPy: Korean natural language processing in Python","volume-title":"Annual Conference on Human and Language Technology. Human and Language Technology","author":"Park","year":"2014"},{"issue":"5","key":"ref16","article-title":"Research on History-based Mongolian Automatic POS Tagging","volume":"27","author":"Jian-dong","year":"2013","journal-title":"Journal of Chinese Information Processing"},{"key":"ref17","first-page":"58","article-title":"An approach based on SV-NN for Kazakh language text classification","volume-title":"Journal of Northeast Normal University","author":"Alimjan","year":"2018"},{"issue":"5","key":"ref18","first-page":"110","article-title":"Machine learning based Uyghur language text categorization","volume":"48","author":"Alimjan","year":"2012","journal-title":"Computer Engineering and Applications"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P16-1162"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-69005-6_39"},{"key":"ref21","first-page":"287","article-title":"An empirical study of topic classification for Korean newspaper headlines","author":"Park","year":"2021","journal-title":"Hum. Lang. Technol"},{"key":"ref22","volume-title":"Convolutional neural network for sentence classification","author":"Chen","year":"2015"},{"key":"ref23","first-page":"2873","article-title":"Recurrent neural network for text clas-sification with multi-task learning","volume-title":"Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI\u2018 16)","author":"Liu","year":"2016"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P16-2034"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v29i1.9513"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/e17-2068"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1052"},{"key":"ref28","first-page":"3937","article-title":"CINO: A Chinese Minority Pre-trained Language Model","volume-title":"Proceedings of the 29th International Conference on Computational Linguistics","author":"Yang","year":"2022"},{"key":"ref29","first-page":"211","article-title":"Tibetan POS Tagging Based on Syllable Tagging","author":"Long","year":"2015","journal-title":"Journal of Chinese Information Processing"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.41"}],"event":{"name":"2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC)","location":"Honolulu, Oahu, HI, USA","start":{"date-parts":[[2023,10,1]]},"end":{"date-parts":[[2023,10,4]]}},"container-title":["2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10391856\/10393862\/10393961.pdf?arnumber=10393961","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,14]],"date-time":"2024-03-14T00:55:00Z","timestamp":1710377700000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10393961\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,1]]},"references-count":30,"URL":"https:\/\/doi.org\/10.1109\/smc53992.2023.10393961","relation":{},"subject":[],"published":{"date-parts":[[2023,10,1]]}}}