{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,19]],"date-time":"2026-04-19T00:47:20Z","timestamp":1776559640607,"version":"3.51.2"},"reference-count":62,"publisher":"Tech Science Press","issue":"1","license":[{"start":{"date-parts":[[2024,10,20]],"date-time":"2024-10-20T00:00:00Z","timestamp":1729382400000},"content-version":"vor","delay-in-days":293,"URL":"https:\/\/doi.org\/10.32604\/TSP-CROSSMARKPOLICY"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["CMC"],"published-print":{"date-parts":[[2024]]},"DOI":"10.32604\/cmc.2024.054673","type":"journal-article","created":{"date-parts":[[2024,9,26]],"date-time":"2024-09-26T07:52:58Z","timestamp":1727337178000},"page":"951-969","update-policy":"https:\/\/doi.org\/10.32604\/tsp-crossmarkpolicy","source":"Crossref","is-referenced-by-count":4,"title":["LKMT: Linguistics Knowledge-Driven Multi-Task Neural Machine Translation for Urdu and English"],"prefix":"10.32604","volume":"81","author":[{"given":"Muhammad Naeem Ul","family":"Hassan","sequence":"first","affiliation":[]},{"given":"Zhengtao","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Ying","family":"Li","sequence":"additional","affiliation":[]},{"given":"Shengxiang","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Shuwan","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Cunli","family":"Mao","sequence":"additional","affiliation":[]}],"member":"17807","published-online":{"date-parts":[[2024]]},"reference":[{"key":"ref1","unstructured":"D. Bahdanau, K. Cho, and Y. Bengio, \u201cNeural machine translation by jointly learning to align and translate,\u201d Sep. 2014, arXiv:1409.0473."},{"key":"ref2","series-title":"In First Conf. Mach. Trans., Assoc. Comput. Linguist.","first-page":"131","article-title":"Findings of the 2016 conference on machine translation (WMT16)","author":"Bojar","year":"2016"},{"key":"ref3","series-title":"Proc. Second Conf. Mach. Trans.","first-page":"169","article-title":"Findings of the 2017 conference on machine translation (WMT17)","author":"Bojar","year":"2017"},{"key":"ref4","series-title":"Proc. Third Conf. Mach. Trans.: Res. Pap.","first-page":"244","article-title":"Trivial transfer learning for low-resource neural machine translation","author":"Kocmi","year":"2018"},{"key":"ref5","unstructured":"Y. Wu et al., \u201cGoogle\u2019s neural machine translation system: Bridging the gap between human and machine translation,\u201d Sep. 2016, arXiv:1609.08144."},{"key":"ref6","unstructured":"H. Hassan et al., \u201cAchieving human parity on automatic chinese to english news translation,\u201d Mar. 2018, arXiv:1803.05567."},{"key":"ref7","first-page":"1723","article-title":"Multi-task learning for multiple language translation","volume":"1","author":"Dong","year":"2015","journal-title":"Proc. 53rd Annu. Meet. Assoc. Comput. Linguist. 7th Int. Joint Conf. Natural Lang. Process."},{"key":"ref8","series-title":"Proc. 2016 Conf. Empir. Methods Natural Lang. Process","first-page":"268","article-title":"Zero-resource translation with multi-lingual neural machine translation","author":"Firat","year":"2016"},{"key":"ref9","unstructured":"T. -L. Ha, J. Niehues, and A. Waibel, \u201cToward multilingual neural machine translation with universal encoder and decoder,\u201d Nov. 2016, arXiv:1611.04798."},{"key":"ref10","doi-asserted-by":"crossref","first-page":"339","DOI":"10.1162\/tacl_a_00065","article-title":"Google\u2019s multilingual neural machine translation system: enabling zero-shot translation","volume":"5","author":"Johnson","year":"2017","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"ref11","doi-asserted-by":"crossref","unstructured":"K. Cho, B. van Merrienboer, D. Bahdanau, and Y. Bengio, \u201cOn the properties of neural machine translation: encoder-decoder approaches,\u201d Sep. 2014, arXiv:1409.1259.","DOI":"10.3115\/v1\/W14-4012"},{"key":"ref12","unstructured":"A. Vaswani et al., \u201cAttention is all you need,\u201d Jun. 2017, arXiv:1706.03762."},{"key":"ref13","first-page":"291","article-title":"Cutting-off redundant repeating generations for neural abstractive summarization","volume":"2","author":"Suzuki","year":"2017","journal-title":"Proc. 15th Conf. Europ. Chapter Assoc. Comput. Linguist."},{"key":"ref14","series-title":"Int. Conf. Mach. Learn.","first-page":"1243","article-title":"Convolutional sequence to sequence learning, Convolutional sequence to sequence learning","author":"Gehring","year":"2017"},{"key":"ref15","first-page":"2773","article-title":"Grammar as a foreign language","volume":"28","author":"Vinyals","year":"2015","journal-title":"Adv. Neural inf. Process. Syst."},{"key":"ref16","unstructured":"J. Devlin, M. -W. Chang, K. Lee, and K. Toutanova, \u201cBERT: Pre-training of deep bidirectional transformers for language understanding,\u201d 2019, arXiv:1810.04805."},{"key":"ref17","unstructured":"Y. Liu et al., \u201cRoBERTa: A robustly optimized BERT pretraining approach,\u201d Jul. 2019, arXiv:1907.11692."},{"key":"ref18","unstructured":"A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, \u201cImproving language understanding by generative pre-training,\u201d 2018. Accessed: May 10, 2024. [Online]. Available: https:\/\/www.cs.ubc.ca\/~amuham01\/LING530\/papers\/radford2018improving.pdf."},{"key":"ref19","unstructured":"Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma and R. Soricut, \u201cALBERT: A lite BERT for self-supervised learning of language representations,\u201d Sep. 2019, arXiv:1909.11942."},{"key":"ref20","doi-asserted-by":"crossref","unstructured":"Z. Zhang, X. Han, Z. Liu, X. Jiang, M. Sun and Q. Liu, \u201cERNIE: enhanced language representation with informative entities,\u201d May 2019, arXiv:1905.07129.","DOI":"10.18653\/v1\/P19-1139"},{"key":"ref21","unstructured":"P. He, X. Liu, J. Gao, and W. Chen, \u201cDeBERTa: Decoding-enhanced BERT with disentangled attention,\u201d Jun. 2020, arXiv:2006.03654."},{"key":"ref22","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3491065","article-title":"Linguistically driven multi-task pre-training for low-resource neural machine translation","volume":"21","author":"Mao","year":"2022","journal-title":"ACM Trans. Asian Low Resour. Lang. Inf. Process."},{"key":"ref23","unstructured":"T. Mikolov, I. Sutskever, K. Chen, G. Corrado, and J. Dean, \u201cDistributed representations of words and phrases and their compositionality,\u201d Oct. 2013, arXiv:1310.4546."},{"key":"ref24","series-title":"Proc. 2014 Conf. Empir. Methods Nat. Lang. Process. (EMNLP)","first-page":"1532","article-title":"GloVe: Global vectors for word representation","author":"Pennington","year":"2014"},{"key":"ref25","first-page":"11468","article-title":"Deep fusing pre-trained models into neural machine translation","volume":"36","author":"Weng","year":"2022","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"ref26","unstructured":"L. Han, G. Erofeev, I. Sorokina, S. Gladkoff, and G. Nenadic, \u201cExamining large pre-trained language models for machine translation: What you don\u2019t know about it,\u201d 2022, arXiv:2209.07417."},{"key":"ref27","doi-asserted-by":"crossref","unstructured":"Z. Sun, M. Wang, and L. Li, \u201cMultilingual translation via grafting pre-trained language models,\u201d 2021, arXiv:2109.05256.","DOI":"10.18653\/v1\/2021.findings-emnlp.233"},{"key":"ref28","series-title":"Proc. Annu. Meet. Assoc. Comput. Linguist.","first-page":"6379","article-title":"Universal conditional masked language pre-training for neural machine translation","author":"Li","year":"2022"},{"key":"ref29","doi-asserted-by":"crossref","unstructured":"J. Hwang and C. -S. Jeong, \u201cIntegrating pre-trained language model into neural machine translation,\u201d Oct. 2023, arXiv:2310.19680.","DOI":"10.1109\/CISDS61173.2023.00019"},{"key":"ref30","doi-asserted-by":"crossref","first-page":"726","DOI":"10.1162\/tacl_a_00343","article-title":"Multilingual denoising pre-training for neural machine translation","volume":"8","author":"Liu","year":"2020","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"ref31","first-page":"9652","article-title":"Reinforced curriculum learning on pre-trained neural machine translation models","volume":"34","author":"Zhao","year":"2020","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"ref32","doi-asserted-by":"crossref","unstructured":"Z. He, G. Blackwood, R. Panda, J. McAuley, and R. Feris, \u201cSynthetic pre-training tasks for neural machine translation,\u201d 2022, arXiv:2212.09864.","DOI":"10.18653\/v1\/2023.findings-acl.512"},{"key":"ref33","first-page":"9266","article-title":"Acquiring knowledge from pre-trained model to neural machine translation","volume":"34","author":"Weng","year":"2020","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"ref34","doi-asserted-by":"crossref","DOI":"10.1155\/2022\/7873012","article-title":"Machine translation system using deep learning for english to Urdu","volume":"2022","author":"Andrabi","year":"2022","journal-title":"Comput. Intell. Neurosci."},{"key":"ref35","doi-asserted-by":"crossref","unstructured":"H. S. Shavarani and A. Sarkar, \u201cBetter neural machine translation by extracting linguistic information from BERT,\u201d 2021, arXiv:2104.02831.","DOI":"10.18653\/v1\/2021.eacl-main.241"},{"key":"ref36","doi-asserted-by":"crossref","first-page":"2341","DOI":"10.1109\/TASLP.2018.2864648","article-title":"Linguistic knowledge-aware neural machine translation","volume":"26","author":"Li","year":"2018","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"ref37","doi-asserted-by":"crossref","first-page":"1047","DOI":"10.1007\/s10772-021-09865-5","article-title":"Linguistically enhanced word segmentation for better neural machine translation of low resource agglutinative languages","volume":"24","author":"Chimalamarri","year":"2021","journal-title":"Int. J. Speech Technol."},{"key":"ref38","doi-asserted-by":"crossref","first-page":"71","DOI":"10.1007\/s10590-021-09263-3","article-title":"Experience of neural machine translation between Indian languages","volume":"35","author":"Dewangan","year":"2021","journal-title":"Mach. Transl."},{"key":"ref39","doi-asserted-by":"crossref","DOI":"10.3390\/fi12060096","article-title":"Multi-source neural model for machine translation of agglutinative language","volume":"12","author":"Pan","year":"2020","journal-title":"Future Internet"},{"key":"ref40","series-title":"13th Int. Conf. Comput. Commun. Netw. Technol. (ICCCNT)","first-page":"1","article-title":"Unsupervised neural machine translation for english to kannada using pre-trained language model","author":"Sheshadri","year":"2022"},{"key":"ref41","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3594631","article-title":"Low-resource multilingual neural translation using linguistic feature-based relevance mechanisms","volume":"22","author":"Chakrabarty","year":"2023","journal-title":"ACM Trans. Asian Low Resour. Lang. Inf. Process."},{"key":"ref42","unstructured":"H. Choudhary, S. Rao, and R. Rohilla, \u201cNeural machine translation for low-resourced Indian languages,\u201d 2020, arXiv:2004.13819."},{"key":"ref43","unstructured":"P. Zaremoodi and G. Haffari, \u201cLearning to multi-task learn for better neural machine translation,\u201d Jan. 2020, arXiv:2001.03294."},{"key":"ref44","doi-asserted-by":"crossref","first-page":"368","DOI":"10.1016\/j.neucom.2021.08.019","article-title":"Improving neural machine translation with latent features feedback","volume":"463","author":"Li","year":"2021","journal-title":"Neurocomputing"},{"key":"ref45","series-title":"Proc. Mass. Multiling. Nat. Lang. Understanding Workshop (MMNLU-22)","first-page":"1","article-title":"Robust domain adaptation for pre-trained multilingual neural machine translation models","author":"Grosso","year":"2022"},{"key":"ref46","doi-asserted-by":"crossref","unstructured":"X. Liu et al., \u201cOn the copying behaviors of pre-training for neural machine translation,\u201d 2021, arXiv:2107.08212.","DOI":"10.18653\/v1\/2021.findings-acl.373"},{"key":"ref47","article-title":"Factors behind the effectiveness of an unsupervised neural machine translation system between Korean and Japanese","volume":"11","author":"Choi","year":"2021","journal-title":"Appl. Sci."},{"key":"ref48","first-page":"9378","article-title":"Towards making the most of BERT in neural machine translation","volume":"34","author":"Yang","year":"2020","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"ref49","doi-asserted-by":"crossref","unstructured":"P. Qi, Y. Zhang, Y. Zhang, J. Bolton, and C. D. Manning, \u201cStanza: A python natural language processing toolkit for many human languages,\u201d Mar. 2020, arXiv:2003.07082.","DOI":"10.18653\/v1\/2020.acl-demos.14"},{"key":"ref50","first-page":"2938","article-title":"A tagged corpus and a tagger for Urdu","volume":"2","author":"Jawaid","year":"2014","journal-title":"LREC"},{"key":"ref51","unstructured":"D. P. Kingma and J. Ba, \u201cAdam: A method for stochastic optimization,\u201d Dec. 2014, arXiv:1412.6980."},{"key":"ref52","doi-asserted-by":"crossref","unstructured":"G. Lample, M. Ott, A. Conneau, L. Denoyer, and M. Ranzato, \u201cPhrase-based & neural unsupervised machine translation,\u201d Apr. 2018, arXiv:1804.07755.","DOI":"10.18653\/v1\/D18-1549"},{"key":"ref53","first-page":"7059","article-title":"Cross-lingual language model pretraining","volume":"32","author":"Conneau","year":"2019","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref54","unstructured":"K. Song, X. Tan, T. Qin, J. Lu, and T. -Y. Liu, \u201cMASS: Masked sequence to sequence pre-training for language generation,\u201d May 2019, arXiv:1905.02450."},{"key":"ref55","series-title":"Proc. Conf. Empir. Methods Natural Lang. Process. (EMNLP)","first-page":"2649","article-title":"Pre-training multilingual neural machine translation by leveraging alignment information","author":"Lin","year":"2020"},{"key":"ref56","series-title":"Proc. Annu. Meet. Assoc. Comput. Linguist.","first-page":"1849","article-title":"IndicBART: A pre-trained model for indic natural language generation","author":"Dabre","year":"2022"},{"key":"ref57","unstructured":"H. Qin et al., \u201cBIBERT: Accurate fully binarized BERT,\u201d 2022, arXiv:2203.06390."},{"key":"ref58","unstructured":"M. R. Costa-juss\u00e0 et al., \u201cNo language left behind: Scaling human-centered machine translation,\u201d 2022, arXiv:2207.04672."},{"key":"ref59","doi-asserted-by":"crossref","first-page":"522","DOI":"10.1162\/tacl_a_00474","article-title":"The Flores-101 evaluation benchmark for low-resource and multilingual machine translation","volume":"10","author":"Goyal","year":"2021","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"ref60","first-page":"2214","article-title":"Parallel data, tools and interfaces in OPUS","volume":"2012","author":"Tiedemann","year":"2012","journal-title":"LREC"},{"key":"ref61","doi-asserted-by":"crossref","first-page":"87","DOI":"10.2478\/v10108-011-0007-0","article-title":"Word-order issues in english-to-urdu statistical machine translation","volume":"95","author":"Jawaid","year":"2011","journal-title":"The Prague Bull. Math. Linguist."},{"key":"ref62","series-title":"Proc. CoNLL, 2018 Shared Task: Multiling. Parsing from Raw Text to Universal Dependencies","first-page":"81","article-title":"An improved neural network model for joint POS tagging and dependency parsing","author":"Nguyen","year":"2018"}],"container-title":["Computers, Materials &amp; Continua"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.techscience.com\/files\/cmc\/2024\/TSP_CMC-81-1\/TSP_CMC_54673\/TSP_CMC_54673.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,7]],"date-time":"2025-03-07T02:34:36Z","timestamp":1741314876000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.techscience.com\/cmc\/v81n1\/58318"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"references-count":62,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2024]]},"published-print":{"date-parts":[[2024]]}},"URL":"https:\/\/doi.org\/10.32604\/cmc.2024.054673","relation":{},"ISSN":["1546-2226"],"issn-type":[{"value":"1546-2226","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"2024-06-04","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-08-26","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-10-15","order":2,"name":"published","label":"Published Online","group":{"name":"publication_history","label":"Publication History"}}]}}