{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T07:10:03Z","timestamp":1755846603075,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":26,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,12,16]],"date-time":"2022-12-16T00:00:00Z","timestamp":1671148800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,12,16]]},"DOI":"10.1145\/3582768.3582791","type":"proceedings-article","created":{"date-parts":[[2023,6,27]],"date-time":"2023-06-27T19:48:32Z","timestamp":1687895312000},"page":"44-48","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Task-specific pre-training improves models for paraphrase generation"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7391-045X","authenticated-orcid":false,"given":"Oleksandr","family":"Skurzhanskyi","sequence":"first","affiliation":[{"name":"Computer Science and Cybernetics \/ Applied Research, Taras Shevchenko National University of Kyiv \/ Grammarly, Ukraine"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5408-5279","authenticated-orcid":false,"given":"Oleksandr","family":"Marchenko","sequence":"additional","affiliation":[{"name":"Computer Science and Cybernetics, Taras Shevchenko National University of Kyiv, Ukraine"}]}],"member":"320","published-online":{"date-parts":[[2023,6,27]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.eacl-main.180"},{"key":"e_1_3_2_1_2_1","volume-title":"Paraphrase generation with latent bag of words. arXiv preprint arXiv:2001.01941","author":"Fu Yao","year":"2020","unstructured":"[2] Yao Fu, Yansong Feng, and John\u00a0P Cunningham. 2020. Paraphrase generation with latent bag of words. arXiv preprint arXiv:2001.01941 (2020)."},{"key":"e_1_3_2_1_3_1","volume-title":"Latent Template Induction with Gumbel-CRFs. arXiv preprint arXiv:2011.14244","author":"Fu Yao","year":"2020","unstructured":"[3] Yao Fu, Chuanqi Tan, Bin Bi, Mosha Chen, Yansong Feng, and Alexander\u00a0M Rush. 2020. Latent Template Induction with Gumbel-CRFs. arXiv preprint arXiv:2011.14244 (2020)."},{"key":"e_1_3_2_1_4_1","volume-title":"International Conference on Machine Learning. PMLR, 1243\u20131252","author":"Gehring Jonas","year":"2017","unstructured":"[4] Jonas Gehring, Michael Auli, David Grangier, Denis Yarats, and Yann\u00a0N Dauphin. 2017. Convolutional sequence to sequence learning. In International Conference on Machine Learning. PMLR, 1243\u20131252."},{"key":"e_1_3_2_1_5_1","volume-title":"Neural syntactic preordering for controlled paraphrase generation. arXiv preprint arXiv:2005.02013","author":"Goyal Tanya","year":"2020","unstructured":"[5] Tanya Goyal and Greg Durrett. 2020. Neural syntactic preordering for controlled paraphrase generation. arXiv preprint arXiv:2005.02013 (2020)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.aiopen.2021.08.002"},{"key":"e_1_3_2_1_7_1","volume-title":"Long short-term memory. Neural computation 9, 8","author":"Hochreiter Sepp","year":"1997","unstructured":"[7] Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural computation 9, 8 (1997), 1735\u20131780."},{"key":"e_1_3_2_1_8_1","volume-title":"Factorising Meaning and Form for Intent-Preserving Paraphrasing. arXiv preprint arXiv:2105.15053","author":"Hosking Tom","year":"2021","unstructured":"[8] Tom Hosking and Mirella Lapata. 2021. Factorising Meaning and Form for Intent-Preserving Paraphrasing. arXiv preprint arXiv:2105.15053 (2021)."},{"key":"e_1_3_2_1_9_1","volume-title":"Deep encoder, shallow decoder: Reevaluating non-autoregressive machine translation. arXiv preprint arXiv:2006.10369","author":"Kasai Jungo","year":"2020","unstructured":"[9] Jungo Kasai, Nikolaos Pappas, Hao Peng, James Cross, and Noah\u00a0A Smith. 2020. Deep encoder, shallow decoder: Reevaluating non-autoregressive machine translation. arXiv preprint arXiv:2006.10369 (2020)."},{"key":"e_1_3_2_1_10_1","volume-title":"Reformulating unsupervised style transfer as paraphrase generation. arXiv preprint arXiv:2010.05700","author":"Krishna Kalpesh","year":"2020","unstructured":"[10] Kalpesh Krishna, John Wieting, and Mohit Iyyer. 2020. Reformulating unsupervised style transfer as paraphrase generation. arXiv preprint arXiv:2010.05700 (2020)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.3115\/1626355.1626389"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33016834"},{"key":"e_1_3_2_1_15_1","volume-title":"Not Rewrite. arXiv preprint arXiv:2005.12592","author":"Omelianchuk Kostiantyn","year":"2020","unstructured":"[15] Kostiantyn Omelianchuk, Vitaliy Atrasevych, Artem Chernodub, and Oleksandr Skurzhanskyi. 2020. GECToR\u2013Grammatical Error Correction: Tag, Not Rewrite. arXiv preprint arXiv:2005.12592 (2020)."},{"key":"e_1_3_2_1_16_1","volume-title":"Proceedings of the 40th annual meeting of the Association for Computational Linguistics. 311\u2013318","author":"Papineni Kishore","year":"2002","unstructured":"[16] Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics. 311\u2013318."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P15-2070"},{"key":"e_1_3_2_1_18_1","volume-title":"A call for clarity in reporting BLEU scores. arXiv preprint arXiv:1804.08771","author":"Post Matt","year":"2018","unstructured":"[18] Matt Post. 2018. A call for clarity in reporting BLEU scores. arXiv preprint arXiv:1804.08771 (2018)."},{"key":"e_1_3_2_1_19_1","volume-title":"Neural paraphrase generation with stacked residual LSTM networks. arXiv preprint arXiv:1610.03098","author":"Prakash Aaditya","year":"2016","unstructured":"[19] Aaditya Prakash, Sadid\u00a0A Hasan, Kathy Lee, Vivek Datla, Ashequl Qadir, Joey Liu, and Oladimeji Farri. 2016. Neural paraphrase generation with stacked residual LSTM networks. arXiv preprint arXiv:1610.03098 (2016)."},{"key":"e_1_3_2_1_20_1","volume-title":"Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers. 223\u2013231","author":"Snover Matthew","year":"2006","unstructured":"[20] Matthew Snover, Bonnie Dorr, Richard Schwartz, Linnea Micciulla, and John Makhoul. 2006. A study of translation edit rate with targeted human annotation. In Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers. 223\u2013231."},{"key":"e_1_3_2_1_21_1","volume-title":"Are Pre-trained Convolutions Better than Pre-trained Transformers?arXiv preprint arXiv:2105.03322","author":"Tay Yi","year":"2021","unstructured":"[21] Yi Tay, Mostafa Dehghani, Jai Gupta, Dara Bahri, Vamsi Aribandi, Zhen Qin, and Donald Metzler. 2021. Are Pre-trained Convolutions Better than Pre-trained Transformers?arXiv preprint arXiv:2105.03322 (2021)."},{"key":"e_1_3_2_1_22_1","unstructured":"[22] Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan\u00a0N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information processing systems. 5998\u20136008."},{"key":"e_1_3_2_1_23_1","volume-title":"Bilateral multi-perspective matching for natural language sentences. arXiv preprint arXiv:1702.03814","author":"Wang Zhiguo","year":"2017","unstructured":"[23] Zhiguo Wang, Wael Hamza, and Radu Florian. 2017. Bilateral multi-perspective matching for natural language sentences. arXiv preprint arXiv:1702.03814 (2017)."},{"key":"e_1_3_2_1_24_1","volume-title":"ParaNMT-50M: Pushing the limits of paraphrastic sentence embeddings with millions of machine translations. arXiv preprint arXiv:1711.05732","author":"Wieting John","year":"2017","unstructured":"[24] John Wieting and Kevin Gimpel. 2017. ParaNMT-50M: Pushing the limits of paraphrastic sentence embeddings with millions of machine translations. arXiv preprint arXiv:1711.05732 (2017)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.5555\/1873738.1873766"},{"key":"e_1_3_2_1_26_1","volume-title":"Improving grammatical error correction via pre-training a copy-augmented architecture with unlabeled data. arXiv preprint arXiv:1903.00138","author":"Zhao Wei","year":"2019","unstructured":"[26] Wei Zhao, Liang Wang, Kewei Shen, Ruoyu Jia, and Jingming Liu. 2019. Improving grammatical error correction via pre-training a copy-augmented architecture with unlabeled data. arXiv preprint arXiv:1903.00138 (2019)."}],"event":{"name":"NLPIR 2022: 2022 6th International Conference on Natural Language Processing and Information Retrieval","acronym":"NLPIR 2022","location":"Bangkok Thailand"},"container-title":["Proceedings of the 2022 6th International Conference on Natural Language Processing and Information Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3582768.3582791","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3582768.3582791","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T06:55:43Z","timestamp":1755845743000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3582768.3582791"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,12,16]]},"references-count":26,"alternative-id":["10.1145\/3582768.3582791","10.1145\/3582768"],"URL":"https:\/\/doi.org\/10.1145\/3582768.3582791","relation":{},"subject":[],"published":{"date-parts":[[2022,12,16]]},"assertion":[{"value":"2023-06-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}