{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,20]],"date-time":"2025-07-20T04:11:31Z","timestamp":1752984691814,"version":"3.37.3"},"reference-count":41,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"DOI":"10.13039\/501100004663","name":"Ministry of Science and Technology of Taiwan through the Young Scholar Fellowship Program","doi-asserted-by":"publisher","award":["MOST 111-2636-E-011-005"],"award-info":[{"award-number":["MOST 111-2636-E-011-005"]}],"id":[{"id":"10.13039\/501100004663","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Ministry of Transportation and Communications in Taiwan","award":["MOTC-CWB-111-E-06"],"award-info":[{"award-number":["MOTC-CWB-111-E-06"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2022]]},"DOI":"10.1109\/access.2022.3207569","type":"journal-article","created":{"date-parts":[[2022,9,16]],"date-time":"2022-09-16T19:35:52Z","timestamp":1663356952000},"page":"99754-99761","source":"Crossref","is-referenced-by-count":5,"title":["Toward Zero-Shot and Zero-Resource Multilingual Question Answering"],"prefix":"10.1109","volume":"10","author":[{"given":"Chia-Chih","family":"Kuo","sequence":"first","affiliation":[{"name":"Computer Science and Information Engineering Department, National Taiwan University of Science and Technology, Taipei, Taiwan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9656-7551","authenticated-orcid":false,"given":"Kuan-Yu","family":"Chen","sequence":"additional","affiliation":[{"name":"Computer Science and Information Engineering Department, National Taiwan University of Science and Technology, Taipei, Taiwan"}]}],"member":"263","reference":[{"volume-title":"Deep Learning","year":"2016","author":"Goodfellow","key":"ref1"},{"key":"ref2","article-title":"Attention is all you need","author":"Vaswani","year":"2017","journal-title":"arXiv:1706.03762"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1810.04805"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1252"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.747"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-demo.28"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.repl4nlp-1.16"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1015"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1607"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.148"},{"key":"ref11","article-title":"Word translation without parallel data","author":"Conneau","year":"2017","journal-title":"arXiv:1710.04087"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1138"},{"key":"ref13","article-title":"A study of cross-lingual ability and language-specific information in multilingual BERT","author":"Liu","year":"2020","journal-title":"arXiv:2004.09205"},{"key":"ref14","article-title":"Looking for clues of language in multilingual BERT to improve cross-lingual generalization","author":"Liu","year":"2020","journal-title":"arXiv:2010.10041"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.42"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.coling-main.105"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00317"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.653"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.5555\/3454287.3455008"},{"key":"ref20","article-title":"HuggingFace\u2019s transformers: State-of-the-art natural language processing","author":"Wolf","year":"2019","journal-title":"arXiv:1910.03771"},{"key":"ref21","article-title":"Decoupled weight decay regularization","author":"Loshchilov","year":"2017","journal-title":"arXiv:1711.05101"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1810.04805"},{"key":"ref23","first-page":"1137","article-title":"A neural probabilistic language model","volume":"3","author":"Bengio","year":"2003","journal-title":"J. Mach. Learn. Res."},{"key":"ref24","first-page":"1","article-title":"Efficient estimation of word representations in vector space","volume-title":"Proc. 1st Int. Conf. Learn. Represent.","author":"Mikolov"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1162"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1202"},{"issue":"8","key":"ref27","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI Blog"},{"key":"ref28","article-title":"XLNet: Generalized autoregressive pretraining for language understanding","author":"Yang","year":"2019","journal-title":"arXiv:1906.08237"},{"key":"ref29","article-title":"RoBERTa: A robustly optimized BERT pretraining approach","author":"Liu","year":"2019","journal-title":"arXiv:1907.11692"},{"key":"ref30","article-title":"ALBERT: A lite BERT for self-supervised learning of language representations","author":"Lan","year":"2019","journal-title":"arXiv:1909.11942"},{"key":"ref31","first-page":"1","article-title":"Word translation without parallel data","volume-title":"Proc. 6th Int. Conf. Learn. Represent.","author":"Conneau"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00288"},{"key":"ref33","first-page":"7057","article-title":"Cross-lingual language model pretraining","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Lample"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.747"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.747"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-2012"},{"key":"ref37","article-title":"InfoXLM: An information-theoretic framework for cross-lingual language model pre-training","author":"Chi","year":"2020","journal-title":"arXiv:2007.07834"},{"key":"ref38","article-title":"ELECTRA: Pre-training text encoders as discriminators rather than generators","author":"Clark","year":"2020","journal-title":"arXiv:2003.10555"},{"key":"ref39","article-title":"XLM-E: Cross-lingual language model pre-training via ELECTRA","author":"Chi","year":"2021","journal-title":"arXiv:2106.16138"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.5555\/3454287.3455008"},{"key":"ref41","first-page":"38","article-title":"Transformers: State-of-the-art natural language processing","volume-title":"Proc. Conf. Empirical Methods Natural Lang. Processing, Syst. Demonstrations","author":"Debut"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/9668973\/09894421.pdf?arnumber=9894421","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T21:08:08Z","timestamp":1705957688000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9894421\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"references-count":41,"URL":"https:\/\/doi.org\/10.1109\/access.2022.3207569","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2022]]}}}