{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T19:50:28Z","timestamp":1743018628489,"version":"3.40.3"},"publisher-location":"Cham","reference-count":27,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031085291"},{"type":"electronic","value":"9783031085307"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-08530-7_31","type":"book-chapter","created":{"date-parts":[[2022,8,29]],"date-time":"2022-08-29T12:13:00Z","timestamp":1661775180000},"page":"370-381","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Machine Reading Comprehension Model for\u00a0Low-Resource Languages and\u00a0Experimenting on\u00a0Vietnamese"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1781-0985","authenticated-orcid":false,"given":"Bach Hoang Tien","family":"Nguyen","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3977-288X","authenticated-orcid":false,"given":"Dung Manh","family":"Nguyen","sequence":"additional","affiliation":[]},{"given":"Trang Thi Thu","family":"Nguyen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,8,30]]},"reference":[{"doi-asserted-by":"crossref","unstructured":"Artetxe, M., Ruder, S., Yogatama, D.: On the cross-lingual transferability of monolingual representations. arXiv preprint arXiv:1910.11856 (2019)","key":"31_CR1","DOI":"10.18653\/v1\/2020.acl-main.421"},{"unstructured":"Brown, T.B., et al.: Language models are few-shot learners. arXiv preprint arXiv:2005.14165 (2020)","key":"31_CR2"},{"doi-asserted-by":"crossref","unstructured":"Conneau, A., et al.: Unsupervised cross-lingual representation learning at scale. arXiv preprint arXiv:1911.02116 (2019)","key":"31_CR3","DOI":"10.18653\/v1\/2020.acl-main.747"},{"doi-asserted-by":"crossref","unstructured":"Cui, Y., Chen, Z., Wei, S., Wang, S., Liu, T., Hu, G.: Attention-over-attention neural networks for reading comprehension. arXiv preprint arXiv:1607.04423 (2016)","key":"31_CR4","DOI":"10.18653\/v1\/P17-1055"},{"unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)","key":"31_CR5"},{"doi-asserted-by":"publisher","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171\u20134186. Association for Computational Linguistics, Minneapolis, Minnesota, June 2019. https:\/\/doi.org\/10.18653\/v1\/N19-1423, https:\/\/aclanthology.org\/N19-1423","key":"31_CR6","DOI":"10.18653\/v1\/N19-1423"},{"unstructured":"Di Gangi, M.A., Cattoni, R., Bentivogli, L., Negri, M., Turchi, M.: Must-C: a multilingual speech translation corpus. In: 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 2012\u20132017. Association for Computational Linguistics (2019)","key":"31_CR7"},{"key":"31_CR8","first-page":"1693","volume":"28","author":"KM Hermann","year":"2015","unstructured":"Hermann, K.M., et al.: Teaching machines to read and comprehend. Adv. Neural. Inf. Process. Syst. 28, 1693\u20131701 (2015)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"doi-asserted-by":"crossref","unstructured":"Hirschman, L., Light, M., Breck, E., Burger, J.D.: Deep read: a reading comprehension system. In: Proceedings of the 37th Annual Meeting of the Association for Computational Linguistics, pp. 325\u2013332 (1999)","key":"31_CR9","DOI":"10.3115\/1034678.1034731"},{"doi-asserted-by":"crossref","unstructured":"Joshi, M., Choi, E., Weld, D.S., Zettlemoyer, L.: Triviaqa: a large scale distantly supervised challenge dataset for reading comprehension. arXiv preprint arXiv:1705.03551 (2017)","key":"31_CR10","DOI":"10.18653\/v1\/P17-1147"},{"doi-asserted-by":"crossref","unstructured":"Lewis, M., et al.: Bart: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461 (2019)","key":"31_CR11","DOI":"10.18653\/v1\/2020.acl-main.703"},{"doi-asserted-by":"crossref","unstructured":"Nguyen, D.Q., Nguyen, A.T.: PhoBert: pre-trained language models for vietnamese. arXiv preprint arXiv:2003.00744 (2020)","key":"31_CR12","DOI":"10.18653\/v1\/2020.findings-emnlp.92"},{"doi-asserted-by":"crossref","unstructured":"Nguyen, K., Nguyen, V., Nguyen, A., Nguyen, N.: A Vietnamese dataset for evaluating machine reading comprehension. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 2595\u20132605 (2020)","key":"31_CR13","DOI":"10.18653\/v1\/2020.coling-main.233"},{"doi-asserted-by":"crossref","unstructured":"Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: Squad: 100,000+ questions for machine comprehension of text. arXiv preprint arXiv:1606.05250 (2016)","key":"31_CR14","DOI":"10.18653\/v1\/D16-1264"},{"unstructured":"Richardson, M., Burges, C.J., Renshaw, E.: MCTest: a challenge dataset for the open-domain machine comprehension of text. In: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pp. 193\u2013203 (2013)","key":"31_CR15"},{"doi-asserted-by":"crossref","unstructured":"Rikters, M., Ri, R., Li, T., Nakazawa, T.: Designing the business conversation corpus. arXiv preprint arXiv:2008.01940 (2020)","key":"31_CR16","DOI":"10.18653\/v1\/D19-5204"},{"doi-asserted-by":"crossref","unstructured":"Riloff, E., Thelen, M.: A rule-based question answering system for reading comprehension tests. In: ANLP-NAACL 2000 Workshop: Reading Comprehension Tests as Evaluation for Computer-Based Language Understanding Systems (2000)","key":"31_CR17","DOI":"10.3115\/1117595.1117598"},{"unstructured":"Seo, M., Kembhavi, A., Farhadi, A., Hajishirzi, H.: Bidirectional attention flow for machine comprehension. arXiv preprint arXiv:1611.01603 (2016)","key":"31_CR18"},{"unstructured":"de Souza, L.R., Nogueira, R., Lotufo, R.: On the ability of monolingual models to learn language-agnostic representations. arXiv preprint arXiv:2109.01942 (2021)","key":"31_CR19"},{"unstructured":"The, V.B., Thi, O.T., Le-Hong, P.: Improving sequence tagging for Vietnamese text using transformer-based neural models. arXiv preprint arXiv:2006.15994 (2020)","key":"31_CR20"},{"doi-asserted-by":"crossref","unstructured":"Tran, N.L., Le, D.M., Nguyen, D.Q.: BartPho: pre-trained sequence-to-sequence models for Vietnamese. arXiv preprint arXiv:2109.09701 (2021)","key":"31_CR21","DOI":"10.21437\/Interspeech.2022-10177"},{"doi-asserted-by":"crossref","unstructured":"Trischler, A., et al.: NewsQA: a machine comprehension dataset. arXiv preprint arXiv:1611.09830 (2016)","key":"31_CR22","DOI":"10.18653\/v1\/W17-2623"},{"key":"31_CR23","doi-asserted-by":"publisher","first-page":"201404","DOI":"10.1109\/ACCESS.2020.3035701","volume":"8","author":"K Van Nguyen","year":"2020","unstructured":"Van Nguyen, K., Tran, K.V., Luu, S.T., Nguyen, A.G.T., Nguyen, N.L.T.: Enhancing lexical-based approach with external knowledge for Vietnamese multiple-choice machine reading comprehension. IEEE Access 8, 201404\u2013201417 (2020)","journal-title":"IEEE Access"},{"doi-asserted-by":"crossref","unstructured":"Van Nguyen, K., Van Huynh, T., Nguyen, D.V., Nguyen, A.G.T., Nguyen, N.L.T.: New Vietnamese corpus for machine reading comprehension of health news articles. arXiv preprint arXiv:2006.11138 (2020)","key":"31_CR24","DOI":"10.18653\/v1\/2020.coling-main.233"},{"unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)","key":"31_CR25"},{"doi-asserted-by":"crossref","unstructured":"Xue, L., et al.: mt5: a massively multilingual pre-trained text-to-text transformer. arXiv preprint arXiv:2010.11934 (2020)","key":"31_CR26","DOI":"10.18653\/v1\/2021.naacl-main.41"},{"doi-asserted-by":"crossref","unstructured":"Zhu, J., Zhou, Y., Zhang, J., Zong, C.: Attend, translate and summarize: an efficient method for neural cross-lingual summarization. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 1309\u20131321 (2020)","key":"31_CR27","DOI":"10.18653\/v1\/2020.acl-main.121"}],"container-title":["Lecture Notes in Computer Science","Advances and Trends in Artificial Intelligence. Theory and Practices in Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-08530-7_31","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T16:02:34Z","timestamp":1710259354000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-08530-7_31"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031085291","9783031085307"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-08530-7_31","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"30 August 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"IEA\/AIE","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Industrial, Engineering and Other Applications of Applied Intelligent Systems","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kitakyushu","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 July 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 July 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"35","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ieaaie2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ieaaie2022.wordpress.com\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"127","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"67","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"14","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"53% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}