{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T14:09:14Z","timestamp":1774534154332,"version":"3.50.1"},"reference-count":199,"publisher":"Association for Computing Machinery (ACM)","issue":"10s","license":[{"start":{"date-parts":[[2022,1,31]],"date-time":"2022-01-31T00:00:00Z","timestamp":1643587200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Comput. Surv."],"published-print":{"date-parts":[[2022,1,31]]},"abstract":"<jats:p>\n            AI is widely thought to be poised to transform business, yet current perceptions of the scope of this transformation may be myopic. Recent progress in natural language processing involving\n            <jats:bold>transformer language models (TLMs)<\/jats:bold>\n            offers a potential avenue for AI-driven business and societal transformation that is beyond the scope of what most currently foresee. We review this recent progress as well as recent literature utilizing text mining in top IS journals to develop an outline for how future IS research can benefit from these new techniques. Our review of existing IS literature reveals that suboptimal text mining techniques are prevalent and that the more advanced TLMs could be applied to enhance and increase IS research involving text data, and to enable new IS research topics, thus creating more value for the research community. This is possible because these techniques make it easier to develop very powerful custom systems and their performance is superior to existing methods for a wide range of tasks and applications. Further, multilingual language models make possible higher quality text analytics for research in multiple languages. We also identify new avenues for IS research, like language user interfaces, that may offer even greater potential for future IS research.\n          <\/jats:p>","DOI":"10.1145\/3505245","type":"journal-article","created":{"date-parts":[[2022,1,5]],"date-time":"2022-01-05T16:25:26Z","timestamp":1641399926000},"page":"1-35","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":47,"title":["Deep Transfer Learning &amp; Beyond: Transformer Language Models in Information Systems Research"],"prefix":"10.1145","volume":"54","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1475-7480","authenticated-orcid":false,"given":"Ross","family":"Gruetzemacher","sequence":"first","affiliation":[{"name":"Wichita State University, W. Frank Barton School of Business"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0287-3249","authenticated-orcid":false,"given":"David","family":"Paradice","sequence":"additional","affiliation":[{"name":"Auburn University, Harbert College of Business"}]}],"member":"320","published-online":{"date-parts":[[2022,9,13]]},"reference":[{"key":"e_1_3_4_2_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2018\/13239"},{"key":"e_1_3_4_3_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2019.0847"},{"key":"e_1_3_4_4_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2017.0768"},{"key":"e_1_3_4_5_1","unstructured":"Ashutosh Adhikari Achyudh Ram Raphael Tang and Jimmy Lin. 2019. DocBERT: BERT for document classification. arXiv:1904.08398. Retrieved from https:\/\/arxiv.org\/abs\/1904.08398."},{"key":"e_1_3_4_6_1","unstructured":"Daniel Adiwardana Minh-Thang Luong David R. So Jamie Hall Noah Fiedel Romal Thoppilan Zi Yang Apoorv Kulshreshtha Gaurav Nemade Yifeng Lu and Quoc V. Le. 2020. Towards a human-like open-domain chatbot. arXiv:2001.09977. Retrieved from https:\/\/arxiv.org\/abs\/2001.09977."},{"key":"e_1_3_4_7_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.468"},{"key":"e_1_3_4_8_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2020\/15379"},{"key":"e_1_3_4_9_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.236"},{"key":"e_1_3_4_10_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2019\/14530"},{"key":"e_1_3_4_11_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1371"},{"key":"e_1_3_4_12_1","volume-title":"Keynote Address at the 12th International Conference on Natural Language Generation","author":"Belz Anya","year":"2019","unstructured":"Anya Belz. 2019. DeepFake news generation: Methods, detection and wider implications. Keynote Address at the 12th International Conference on Natural Language Generation. Tokyo, Japan (Oct. 2019)."},{"key":"e_1_3_4_13_1","doi-asserted-by":"publisher","DOI":"10.5555\/944919.944966"},{"key":"e_1_3_4_14_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2016.1205918"},{"key":"e_1_3_4_15_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2019\/13808"},{"key":"e_1_3_4_16_1","first-page":"1533","volume-title":"Proceedings of Empirical Methods in Natural Language Processing 2013. ACL","author":"Berant Jonathan","year":"2013","unstructured":"Jonathan Berant, Andrew Chou, Roy Frostig, and Percy Liang. 2013. Semantic parsing on freebase from question-answer pairs. In Proceedings of Empirical Methods in Natural Language Processing 2013. ACL, 1533\u20131544."},{"key":"e_1_3_4_17_1","first-page":"32","volume-title":"Proceedings of First Workshop on Insights from Negative Results in NLP. ACL","author":"Bianchi Federico","year":"2020","unstructured":"Federico Bianchi, Silvia Terragni, and Dirk Hovy. 2020. Pre-training is a hot topic: Contextualized document embeddings improve topic coherence. In Proceedings of First Workshop on Insights from Negative Results in NLP. ACL, 32\u201340."},{"key":"e_1_3_4_18_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.703"},{"key":"e_1_3_4_19_1","doi-asserted-by":"publisher","DOI":"10.5555\/944919.944937"},{"key":"e_1_3_4_20_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2015.0605"},{"key":"e_1_3_4_21_1","unstructured":"Rishi Bommasani Drew A. Hudso Ehsan Adeli Russ Altman Simran Arora Sydney von Arx Michael S. Bernstein Jeannette Bohg Antoine Bosselut Emma Brunskill Erik Brynjolfsson Shyamal Buch Dallas Card Rodrigo Castellon Niladri Chatterji Annie Chen Kathleen Creel Jared Quincy Davis Dorottya Demszky Chris Donahue Moussa Doumbouya Esin Durmus Stefano Ermon John Etchemendy Kawin Ethayarajh Li Fei-Fei Chelsea Finn Trevor Gale Lauren Gillespie Karan Goel Noah Goodman Shelby Grossman Neel Guha Tatsunori Hashimoto Peter Henderson John Hewitt Daniel E. Ho Jenny Hong Kyle Hsu Jing Huang Thomas Icard Saahil Jain Dan Jurafsky Pratyusha Kalluri Siddharth Karamcheti Geoff Keeling Fereshte Khani Omar Khattab Pang Wei Koh Mark Krass Ranjay Krishna Rohith Kuditipudi Ananya Kumar Faisal Ladhak Mina Lee Tony Lee Jure Leskovec Isabelle Levent Xiang Lisa Li Xuechen Li Tengyu Ma Ali Malik Christopher D. Manning Suvir Mirchandani Eric Mitchell Zanele Munyikwa Suraj Nair Avanika Narayan Deepak Narayanan Ben Newman Allen Nie Juan Carlos Niebles Hamed Nilforoshan Julian Nyarko Giray Ogut Laurel Orr Isabel Papadimitriou Joon Sung Park Chris Piech Eva Portelance Christopher Potts Aditi Raghunathan Rob Reich Hongyu Ren Frieda Rong Yusuf Roohani Camilo Ruiz Jack Ryan Christopher R\u00e9 Dorsa Sadigh Shiori Sagawa Keshav Santhanam Andy Shih Krishnan Srinivasan Alex Tamkin Rohan Taori Armin W. Thomas Florian Tram\u00e8r Rose E. Wang William Wang Bohan Wu Jiajun Wu Yuhuai Wu Sang Michael Xie Michihiro Yasunaga Jiaxuan You Matei Zaharia Michael Zhang Tianyi Zhang Xikun Zhang Yuhui Zhang Lucia Zheng Kaitlyn Zhou and Percy Liang. 2021. On the Opportunities and Risks of Foundation Models. arXiv:2108.07258. Retrieved from https:\/\/arxiv.org\/abs\/2108.07258."},{"key":"e_1_3_4_22_1","doi-asserted-by":"publisher","DOI":"10.1007\/BF00158952"},{"key":"e_1_3_4_23_1","volume-title":"Advances in Neural Information Processing Systems","author":"Brown Tom B.","year":"2020","unstructured":"Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language models are few-shot learners. In Advances in Neural Information Processing Systems 33 (NeurIPS 2020)."},{"key":"e_1_3_4_24_1","volume-title":"Understanding and Addressing the Modern Productivity Paradox","author":"Brynjolfsson Erik","year":"2020","unstructured":"Erik Brynjolfsson, Seth Benzell, and Daniel Rock. 2020. Understanding and Addressing the Modern Productivity Paradox. Research Brief, MIT Work of the Future Task Force. Massachusetts Institute of Technology, Cambrdige, MA."},{"key":"e_1_3_4_25_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.554"},{"key":"e_1_3_4_26_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/S19-2005"},{"key":"e_1_3_4_27_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chb.2018.12.029"},{"key":"e_1_3_4_28_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2020\/14110"},{"key":"e_1_3_4_29_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.152"},{"key":"e_1_3_4_30_1","unstructured":"Brian Cheang Bailey Wei David Kogan Howey Qiu and Masud Ahmed. 2020. Language representation models for fine-grained sentiment classification. arXiv:2005.13619. Retrieved from https:\/\/arxiv.org\/abs\/2005.13619."},{"key":"e_1_3_4_31_1","doi-asserted-by":"publisher","DOI":"10.5555\/2481674.2481683"},{"key":"e_1_3_4_32_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2020.0969"},{"key":"e_1_3_4_33_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2018.1550547"},{"key":"e_1_3_4_34_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2018.0781"},{"key":"e_1_3_4_35_1","doi-asserted-by":"publisher","DOI":"10.1007\/s41019-019-0087-7"},{"key":"e_1_3_4_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401044"},{"key":"e_1_3_4_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3193120"},{"key":"e_1_3_4_38_1","volume-title":"Proceedings of the International Conference on Learning Representations 2021.","author":"Choromanski Krzysztof","year":"2021","unstructured":"Krzysztof Choromanski, Valerii Likhosherstov, David Dohan, Xingyou Song, Andreea Gane, Tamas Sarlos, Peter Hawkins, Jared Davis, Afroz Mohiuddin, Lukasz Kaiser, David Belanger, Lucy Colwell, and Adrian Weller. 2021. Rethinking attention with performers. In Proceedings of the International Conference on Learning Representations 2021."},{"key":"e_1_3_4_39_1","volume-title":"Proceedings of NIPS 2014 Workshop on Deep Learning.","author":"Chung Junyoung","year":"2014","unstructured":"Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, and Yoshua Bengio. 2014. Empirical evaluation of gated recurrent neural networks on sequence modeling. In Proceedings of NIPS 2014 Workshop on Deep Learning."},{"key":"e_1_3_4_40_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2019.0884"},{"key":"e_1_3_4_41_1","doi-asserted-by":"crossref","unstructured":"Peter Clark Oren Etzioni Daniel Khashabi Tushar Khot Bhavana Dalvi Mishra Kyle Richardson Ashish Sabharwal Carissa Schoenick Oyvind Tafjord Niket Tandon Sumithra Bhakthavatsalam Dirk Groeneveld Michal Guerquin and Michael Schmitz. 2019. From 'F' to 'A' on the NY Regents science exams: An overview of the Aristo Project. arXiv:1909.01958. Retrieved from https:\/\/arxiv.org\/abs\/1909.01958","DOI":"10.1609\/aimag.v41i4.5304"},{"key":"e_1_3_4_42_1","doi-asserted-by":"publisher","DOI":"10.5555\/1953048.2078186"},{"key":"e_1_3_4_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/1390156.1390177"},{"key":"e_1_3_4_44_1","first-page":"8440","volume-title":"Proceedings of Annual Meeting of the Association of Computational Linguistics 2020. ACL","author":"Conneau Alexis","year":"2019","unstructured":"Alexis Conneau, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzm\u00e1n, Edouard Grave, Myle Ott, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Unsupervised cross-lingual representation learning at scale. In Proceedings of Annual Meeting of the Association of Computational Linguistics 2020. ACL, 8440\u20138451."},{"key":"e_1_3_4_45_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jfineco.2016.05.001"},{"key":"e_1_3_4_46_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1285"},{"key":"e_1_3_4_47_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2018\/14268"},{"key":"e_1_3_4_48_1","unstructured":"Harm de Vries Dzmitry Bahdanau and Christopher Manning. 2020. Towards ecologically valid research on language user interfaces. arXiv:2007.14435. Retrieved from https:\/\/arxiv.org\/abs\/2007.14435."},{"key":"e_1_3_4_49_1","first-page":"4171","volume-title":"Proceedings of the Annual Meeting of the North American Chapter of the Association of Computational Linguistics 2019. ACL","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the Annual Meeting of the North American Chapter of the Association of Computational Linguistics 2019. ACL, 4171\u20134186."},{"key":"e_1_3_4_50_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2018.1451954"},{"key":"e_1_3_4_51_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1301"},{"key":"e_1_3_4_52_1","article-title":"Cross-lingual cybersecurity analytics in the international dark web with adversarial deep representation learning","author":"Ebrahimi Mohammadreza","year":"2021","unstructured":"Mohammadreza Ebrahimi, Yidong Chai, Sagar Samtani, and Hsinchun Chen. 2021. Cross-lingual cybersecurity analytics in the international dark web with adversarial deep representation learning. MIS Quarterly forthcoming.","journal-title":"MIS Quarterly"},{"issue":"4","key":"e_1_3_4_53_1","first-page":"62","article-title":"Building the AI-powered organization","volume":"97","author":"Fountaine Tim","year":"2019","unstructured":"Tim Fountaine, Brian McCarthy, and Tamim Saleh. 2019. Building the AI-powered organization. Harvard Business Review 97, 4 (2019), 62\u201373.","journal-title":"Harvard Business Review"},{"key":"e_1_3_4_54_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2018\/13215"},{"key":"e_1_3_4_55_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2016.1267526"},{"key":"e_1_3_4_56_1","doi-asserted-by":"crossref","unstructured":"Deepanway Ghosal Navonil Majumder Alexander Gelbukh Rada Mihalcea and Soujanya Poria. 2020. COSMIC: COmmonSense knowledge for eMotion identification in conversations. arXiv:2010.02795. Retrieved from https:\/\/arxiv.org\/abs\/2010.02795.","DOI":"10.18653\/v1\/2020.findings-emnlp.224"},{"key":"e_1_3_4_57_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2018\/14042"},{"key":"e_1_3_4_58_1","doi-asserted-by":"publisher","DOI":"10.1093\/bib\/bbv087"},{"key":"e_1_3_4_59_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.techfore.2020.120323"},{"key":"e_1_3_4_60_1","doi-asserted-by":"publisher","DOI":"10.5121\/ijaia.2019.10203"},{"key":"e_1_3_4_61_1","first-page":"901","volume-title":"Proceedings of 12th Language Resources and Evaluation Conference","author":"Gyawali Bikash","year":"2020","unstructured":"Bikash Gyawali, Lucas Anastasiou, and Petr Knoth. 2020. Deduplication of scholarly documents using locality sensitive hashing and word embeddings. In Proceedings of 12th Language Resources and Evaluation Conference. ACL, 901\u2013910."},{"key":"e_1_3_4_62_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1082"},{"key":"e_1_3_4_63_1","volume-title":"Proceedings of the International Conference on Learning Representations 2021.","author":"He Pengcheng","year":"2021","unstructured":"Pengcheng He, Xiaodong Liu, Jianfeng Gao, and Weizhu Chen. 2021. DeBERTa: Decoding-enhanced BERT with disentangled attention. In Proceedings of the International Conference on Learning Representations 2021."},{"key":"e_1_3_4_64_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2017.0738"},{"key":"e_1_3_4_65_1","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_3_4_66_1","first-page":"2790","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Houlsby Neil","year":"2019","unstructured":"Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin De Laroussilhe, Andrea Gesmundo, Mona Attariyan, and Sylvain Gelly. 2019. Parameter-efficient transfer learning for NLP. In Proceedings of the International Conference on Machine Learning 2019. 2790\u20132799."},{"key":"e_1_3_4_67_1","doi-asserted-by":"crossref","unstructured":"Jeremy Howard and Sebastian Ruder. 2018. Universal language model fine-tuning for text classification. arXiv:1801.06146. Retrieved from https:\/\/arxiv.org\/abs\/1801.06146.","DOI":"10.18653\/v1\/P18-1031"},{"key":"e_1_3_4_68_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.137"},{"key":"e_1_3_4_69_1","doi-asserted-by":"crossref","unstructured":"Jiaxin Huang Yu Meng Fang Guo Heng Ji and Jiawei Han. 2020. Weakly-supervised aspect-based sentiment analysis via joint aspect-sentiment topic embedding. arXiv:2010.06705. https:\/\/arxiv.org\/abs\/2010.06705.","DOI":"10.18653\/v1\/2020.emnlp-main.568"},{"key":"e_1_3_4_70_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2017.1394079"},{"key":"e_1_3_4_71_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2019\/13225"},{"key":"e_1_3_4_72_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2017\/41.4.02"},{"key":"e_1_3_4_73_1","first-page":"65","volume-title":"Proceedings of Empirical Methods in Natural Language Processing","author":"Huang Po-Sen","year":"2019","unstructured":"Po-Sen Huang, Huan Zhang, Ray Jiang, Robert Stanforth, Johannes Welbl, Jack Rae, Vishal Maini, Dani Yogatama, and Pushmeet Kohli. 2019. Reducing sentiment bias in language models via counterfactual evaluation. In Proceedings of Empirical Methods in Natural Language Processing 2019. ACL, 65\u201383."},{"key":"e_1_3_4_74_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2018.0804"},{"key":"e_1_3_4_75_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jfineco.2018.10.001"},{"key":"e_1_3_4_76_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.556"},{"key":"e_1_3_4_77_1","unstructured":"Jared Kaplan Sam McCandlish Tom Henighan Tom B. Brown Benjamin Chess Rewon Child Scott Gray Alec Radford Jeffrey Wu and Dario Amodei. 2020. Scaling laws for neural language models. arXiv:2001.08361. Retrieved from https:\/\/arxiv.org\/abs\/2001.08361."},{"key":"e_1_3_4_78_1","first-page":"6975","volume-title":"Proceedings of Empirical Methods in Natural Language Processing","author":"Ke Pei","year":"2020","unstructured":"Pei Ke, Haozhe Ji, Siyang Liu, Xiaoyan Zhu, and Minlie Huang. 2020. SentiLARE: Linguistic knowledge enhanced language representation for sentiment analysis. In Proceedings of Empirical Methods in Natural Language Processing 2020. ACL, 6975\u20136988."},{"key":"e_1_3_4_79_1","unstructured":"Nitish Shirish Keskar Bryan McCann Lav R. Varshney Caiming Xiong and Richard Socher. 2019. CTRL: A conditional transformer language model for controllable generation. arXiv:1909.05858. Retrieved from https:\/\/arxiv.org\/abs\/1909.05858."},{"key":"e_1_3_4_80_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.171"},{"key":"e_1_3_4_81_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2017.0750"},{"key":"e_1_3_4_82_1","volume-title":"Proceedings of the International Conference on Learning Representations 2020.","author":"Kitaev Nikita","year":"2020","unstructured":"Nikita Kitaev, \u0141ukasz Kaiser, and Anselm Levskaya. 2020. Reformer: The efficient transformer. In Proceedings of the International Conference on Learning Representations 2020."},{"key":"e_1_3_4_83_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.dss.2017.10.001"},{"key":"e_1_3_4_84_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ejor.2019.09.018"},{"key":"e_1_3_4_85_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1051"},{"key":"e_1_3_4_86_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2019.1661089"},{"key":"e_1_3_4_87_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2016.0674"},{"key":"e_1_3_4_88_1","volume-title":"Proceedings of the International Conference on Learning Representations 2019","author":"Lan Zhenzhong","year":"2019","unstructured":"Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2019. ALBERT: A lite BERT for self-supervised learning of language representations. In Proceedings of the International Conference on Learning Representations 2019."},{"key":"e_1_3_4_89_1","doi-asserted-by":"publisher","DOI":"10.1038\/nature14539"},{"key":"e_1_3_4_90_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2020.0921"},{"key":"e_1_3_4_91_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"e_1_3_4_92_1","volume-title":"Advances in Neural Information Processing Systems","author":"Lewis Mike","year":"2020","unstructured":"Mike Lewis, Marjan Ghazvininejad, Gargi Ghosh, Armen Aghajanyan, Sida Wang, and Luke Zettlemoyer. 2020b. Pre-training via paraphrasing. In Advances in Neural Information Processing Systems 33 (NeurIPS 2020)."},{"key":"e_1_3_4_93_1","volume-title":"Advances in Neural Information Processing Systems","volume":"33","author":"Lewis Patrick","year":"2020","unstructured":"Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich K\u00fcttler, Mike Lewis, Wen-tau Yih, Tim Rockt\u00e4schel, Sebastian Riedel, and Douwe Kiela. 2020c. Retrieval-augmented generation for knowledge-intensive NLP tasks. In Advances in Neural Information Processing Systems 33 (NeurIPS 2020)."},{"key":"e_1_3_4_94_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2020\/15323"},{"key":"e_1_3_4_95_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00521-020-05411-7"},{"key":"e_1_3_4_96_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2016.1267528"},{"key":"e_1_3_4_97_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-5505"},{"key":"e_1_3_4_98_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2019\/14360"},{"key":"e_1_3_4_99_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.484"},{"key":"e_1_3_4_100_1","unstructured":"Opher Lieber Or Sharir Barak Lenz and Yoav Shoham. 2021. Jurrasic-1: Technical details and evaluation. White Paper AI21 Labs . Tel Aviv Israel."},{"key":"e_1_3_4_101_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2020\/15107"},{"key":"e_1_3_4_102_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1441"},{"key":"e_1_3_4_103_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2019.0911"},{"key":"e_1_3_4_104_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1387"},{"key":"e_1_3_4_105_1","unstructured":"Yinhan Liu Myle Ott Naman Goyal Jingfei Du Mandar Joshi Danqi Chen Omer Levy Mike Lewis Luke Zettlemoyer and Veselin Stoyanov. 2019c. RoBERTa: A robustly optimized BERT pretraining approach. arXiv:1907.11692 . Retrieved from https:\/\/arxiv.org\/abs\/1907.11692."},{"key":"e_1_3_4_106_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2020.0954"},{"key":"e_1_3_4_107_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.648"},{"key":"e_1_3_4_108_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.602"},{"key":"e_1_3_4_109_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2018.1440774"},{"key":"e_1_3_4_110_1","doi-asserted-by":"publisher","DOI":"10.5555\/311445"},{"key":"e_1_3_4_111_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.173"},{"key":"e_1_3_4_112_1","article-title":"Learned in translation: Contextualized word vectors","volume":"30","author":"McCann Bryan","year":"2017","unstructured":"Bryan McCann, James Bradbury, Caiming Xiong, and Richard Socher. 2017. Learned in translation: Contextualized word vectors. In Advances in Neural Information Processing Systems 30 (NIPS 2017).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_4_113_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2019.0866"},{"key":"e_1_3_4_114_1","first-page":"3111","article-title":"Distributed representations of words and phrases and their compositionality","volume":"26","author":"Mikolov Tomas","year":"2013","unstructured":"Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S. Corrado, and Jeff Dean. 2013a. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems 26 (NIPS 2013). 3111\u20133119.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_4_115_1","volume-title":"Proceedings of the International Conference on Learning Representations 2013.","author":"Mikolov Tomas","year":"2013","unstructured":"Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013b. Efficient estimation of word representations in vector space. In Proceedings of the International Conference on Learning Representations 2013."},{"key":"e_1_3_4_116_1","unstructured":"Tomas Mikolov Armand Joulin Sumit Chopra Michael Mathieu and Marc'Aurelio Ranzato. 2014. Learning longer memory in recurrent neural networks. arXiv:1412.7753. Retrieved from https:\/\/arxiv.org\/abs\/1412.7753."},{"key":"e_1_3_4_117_1","doi-asserted-by":"publisher","DOI":"10.9781\/ijimai.2016.369"},{"key":"e_1_3_4_118_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1900949116"},{"key":"e_1_3_4_119_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2018.0791"},{"key":"e_1_3_4_120_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2019.0889"},{"key":"e_1_3_4_121_1","unstructured":"Martin M\u00fcller Marcel Salath\u00e9 and Per E. Kummervold. 2020. COVID-Twitter-BERT: A natural language processing model to analyse COVID-19 content on Twitter. arXiv:2005.07503. Retrieved from https:\/\/arxiv.org\/abs\/2005.07503."},{"key":"e_1_3_4_122_1","volume-title":"Machine Learning: A Probabilistic Perspective","author":"Murphy Kevin P.","year":"2012","unstructured":"Kevin P. Murphy. 2012. Machine Learning: A Probabilistic Perspective. MIT Press, Cambridge, MA."},{"key":"e_1_3_4_123_1","doi-asserted-by":"publisher","DOI":"10.1613\/jair.1.11315"},{"key":"e_1_3_4_124_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.future.2020.06.050"},{"key":"e_1_3_4_125_1","volume-title":"Proceedings of the Pacific Asia Conference on Information Systems. AIS.","author":"Ngai Eric W. T.","year":"2016","unstructured":"Eric W. T. Ngai and Philip Tin Yun Lee. 2016. A review of the literature on applications of text mining in policy making. In Proceedings of the Pacific Asia Conference on Information Systems. AIS."},{"key":"e_1_3_4_126_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-demos.2"},{"key":"e_1_3_4_127_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2018.0816"},{"key":"e_1_3_4_128_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2019.06.068"},{"key":"e_1_3_4_129_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1162"},{"key":"e_1_3_4_130_1","first-page":"2227","volume-title":"Proceedings of the Annual Meeting of the North American Chapter of the Association of Computational Linguistics","author":"Peters Matthew E.","year":"2018","unstructured":"Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. 2018. Deep contextualized word representations. In Proceedings of the Annual Meeting of the North American Chapter of the Association of Computational Linguistics 2018. ACL, 2227\u20132237."},{"key":"e_1_3_4_131_1","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. (11 2018). Retrieved Dec. 11 2021 from https:\/\/s3-us-west-2.amazonaws.com\/openai-assets\/research-covers\/language-unsupervised\/language_understanding_paper.pdf."},{"key":"e_1_3_4_132_1","unstructured":"Alec Radford Jeffrey Wu Rewon Child David Luan Dario Amodei and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. (14 2019). Retrieved Dec. 11 2021 from https:\/\/d4mucfpksywv.cloudfront.net\/better-language-models\/language_models_are_unsupervised_multitask_learners.pdf."},{"key":"e_1_3_4_133_1","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research 21 (2020), 1\u201367.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_4_134_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1054"},{"issue":"4","key":"e_1_3_4_135_1","first-page":"1127","article-title":"Playing to the crowd? Digital visibility and the social dynamics of purchase disclosure","volume":"43","author":"Rhue Lauren","year":"2019","unstructured":"Lauren Rhue and Arun Sundararajan. 2019. Playing to the crowd? Digital visibility and the social dynamics of purchase disclosure. MIS Quarterly 43, 4 (2019), 1127\u20131141.","journal-title":"MIS Quarterly"},{"key":"e_1_3_4_136_1","doi-asserted-by":"publisher","DOI":"10.1177\/0022243719892594"},{"key":"e_1_3_4_137_1","first-page":"300","volume-title":"Proceedings of the 2020 Conference of the European Chapter of the Association for Computational Linguistics","author":"Roller Stephen","year":"2020","unstructured":"Stephen Roller, Emily Dinan, Naman Goyal, Da Ju, Mary Williamson, Yinhan Liu, Jing Xu, Myle Ott, Kurt Shuster, Eric M. Smith, Y-Lan Boureau, and Jason Weston. 2020a. Recipes for building an open-domain chatbot. In Proceedings of the 2020 Conference of the European Chapter of the Association for Computational Linguistics. ACL, 300\u2013325."},{"key":"e_1_3_4_138_1","unstructured":"Stephen Roller Y-Lan Boureau Jason Weston Antoine Bordes Emily Dinan Angela Fan David Gunning Da Ju Margaret Li Spencer Poff Pratik Ringshia Kurt Shuster Eric Michael Smith Arthur Szlam Jack Urbanek and Mary Williamson. 2020b. Open-domain conversational agents: Current progress open problems and future directions. arXiv:2006.12442 . Retrieved from https:\/\/arxiv.org\/abs\/2006.12442."},{"key":"e_1_3_4_139_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2019.1661093"},{"key":"e_1_3_4_140_1","first-page":"554","volume-title":"Proceedings of the Annual Meeting of the North American Chapter of the Association of Computational Linguistics","author":"Salant Shimi","year":"2018","unstructured":"Shimi Salant and Jonathan Berant. 2018. Contextualized word representations for reading comprehension. In Proceedings of the Annual Meeting of the North American Chapter of the Association of Computational Linguistics. ACL, 554\u2013559."},{"key":"e_1_3_4_141_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2020\/15610"},{"key":"e_1_3_4_142_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2017.1394049"},{"key":"e_1_3_4_143_1","unstructured":"Sagar Samtani Hongyi Zhu Balaji Padmanabhan Yidong Chai and Hsinchun Chen. 2020. Deep learning for information systems research. arXiv:2010.05774. Retrieved from https:\/\/arxiv.org\/abs\/2010.05774."},{"key":"e_1_3_4_144_1","article-title":"Linking exploits from the dark web to known vulnerabilities for proactive cyber threat intelligence: An attention-based deep structured semantic model","author":"Samtani Sagar","year":"2021","unstructured":"Sagar Samtani, Yidong Chai, and Hsinchun Chen. 2021. Linking exploits from the dark web to known vulnerabilities for proactive cyber threat intelligence: An attention-based deep structured semantic model. MIS Quarterly forthcoming.","journal-title":"MIS Quarterly"},{"key":"e_1_3_4_145_1","volume-title":"Proceedings of the 2019 NeurIPS Workshop on Energy Efficient Machine Learning and Cognitive Computing","author":"Sanh Victor","year":"2019","unstructured":"Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. 2019. DistilBERT, a distilled version of BERT: Smaller, faster, cheaper and lighter. In Proceedings of the 2019 NeurIPS Workshop on Energy Efficient Machine Learning and Cognitive Computing."},{"key":"e_1_3_4_146_1","doi-asserted-by":"crossref","unstructured":"Timo Schick and Hinrich Sch\u00fctze. 2020. It's not just size that matters: Small language models are also few-shot learners. arXiv:2009.07118 . Retrieved from https:\/\/arxiv.org\/abs\/2009.07118.","DOI":"10.18653\/v1\/2021.naacl-main.185"},{"key":"e_1_3_4_147_1","article-title":"Inducing brain-relevant bias in natural language processing models","volume":"32","author":"Schwartz Dan","year":"2019","unstructured":"Dan Schwartz, Mariya Toneva, and Leila Wehbe. 2019. Inducing brain-relevant bias in natural language processing models. In Advances in Neural Information Processing Systems 32 (NeurIPS 2019).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_4_148_1","first-page":"8051","volume-title":"Proceedings of Empirical Methods in Natural Language Processing","author":"Scialom Thomas","year":"2020","unstructured":"Thomas Scialom, Paul-Alexis Dray, Sylvain Lamprier, Benjamin Piwowarski, and Jacopo Staiano. 2020. MLSUM: The multilingual summarization corpus. In Proceedings of Empirical Methods in Natural Language Processing 2020. ACL, 8051\u20138067."},{"key":"e_1_3_4_149_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2018.2837384"},{"key":"e_1_3_4_150_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2017.1394056"},{"key":"e_1_3_4_151_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2016\/40.4.11"},{"key":"e_1_3_4_152_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2020\/14870"},{"key":"e_1_3_4_153_1","first-page":"1728","volume-title":"Clusters of pretrained word embeddings make for fast and good topics too! In Proceedings of Empirical Methods in Natural Language Processing 2020","author":"Sia Suzanna","year":"2020","unstructured":"Suzanna Sia, Ayush Dalmia, and Sabrina J. Mielke. 2020. Tired of topic models? Clusters of pretrained word embeddings make for fast and good topics too! In Proceedings of Empirical Methods in Natural Language Processing 2020. ACL, 1728\u20131736."},{"key":"e_1_3_4_154_1","first-page":"4411","volume-title":"Proceedings of the International Conference on Machine Learning 2020","author":"Siddhant Aditya","year":"2020","unstructured":"Aditya Siddhant, Junjie Hu, Melvin Johnson, Orhan Firat, and Sebastian Ruder. 2020. Xtreme: A massively multilingual multi-task benchmark for evaluating cross-lingual generalization. In Proceedings of the International Conference on Machine Learning 2020. 4411\u20134421."},{"key":"e_1_3_4_155_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2016.1205930"},{"key":"e_1_3_4_156_1","doi-asserted-by":"publisher","DOI":"10.1038\/nature24270"},{"key":"e_1_3_4_157_1","doi-asserted-by":"publisher","DOI":"10.1145\/3347145"},{"key":"e_1_3_4_158_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2018.0797"},{"key":"e_1_3_4_159_1","article-title":"Learning to summarize with human feedback","volume":"33","author":"Stiennon Nisan","year":"2020","unstructured":"Nisan Stiennon, Long Ouyang, Jeff Wu, Daniel M. Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, and Paul Christiano. 2020. Learning to summarize with human feedback. In Advances in Neural Information Processing Systems 33 (NeurIPS 2020).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_4_160_1","doi-asserted-by":"crossref","first-page":"194","DOI":"10.1007\/978-3-030-32381-3_16","volume-title":"Proceedings of China National Conference on Chinese Computational Linguistics","author":"Sun Chi","year":"2019","unstructured":"Chi Sun, Xipeng Qiu, Yige Xu, and Xuanjing Huang. 2019. How to fine-tune BERT for text classification?. In Proceedings of China National Conference on Chinese Computational Linguistics. Springer, 194\u2013206."},{"key":"e_1_3_4_161_1","unstructured":"Yu Sun Shuohuan Wang Shikun Feng Siyu Din Chao Pan Junyuan Shan Jiaxiang Li Xuyi Chen Yanbin Zhao Yuxiang Lu Weixin Liu Zhihua Wu Weibao Gong Jianzhong Liang Zhizhou Shang Peng Sun Wei Liu Xuan Ouyang Dianhai Yu Hao Tian Hua Wu and Haifeng Wang. 2021. ERNIE 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation. arXiv:2107.02137 . Retrieved from https:\/\/arxiv.org\/abs\/2107.02137."},{"key":"e_1_3_4_162_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1514"},{"key":"e_1_3_4_163_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.162"},{"key":"e_1_3_4_164_1","volume-title":"Proceedings of the Annual Meeting of the Association of Computational Linguistics 2021.","author":"Tay Yi","year":"2021","unstructured":"Yi Tay, Mostafa Dehghani, Jai Gupta, Dara Bahri, Vamsi Aribandi, Zhen Qin, and Donald Metzler. 2021. Are pre-trained convolutions better than pre-trained transformers? In Proceedings of the Annual Meeting of the Association of Computational Linguistics 2021."},{"key":"e_1_3_4_165_1","first-page":"236","article-title":"Computing machinery and intelligence","volume":"59","author":"Turing Alan. M.","year":"1950","unstructured":"Alan. M. Turing. 1950. Computing machinery and intelligence. Mind 59, 236 (1950), 433\u2013460.","journal-title":"Mind"},{"key":"e_1_3_4_166_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2018.1451961"},{"key":"e_1_3_4_167_1","article-title":"Attention is all you need","volume":"30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems 30 (NIPS 2017).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_4_168_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2019.1598692"},{"key":"e_1_3_4_169_1","first-page":"353","volume-title":"Proceedings of EMNLP Workshop on BlackBox NLP","author":"Wang Alex","year":"2018","unstructured":"Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R. Bowman. 2018a. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In Proceedings of EMNLP Workshop on BlackBox NLP, ACL, 353\u2013355."},{"key":"e_1_3_4_170_1","article-title":"SuperGLUE: A stickier benchmark for general-purpose language understanding systems","volume":"32","author":"Wang Alex","year":"2019","unstructured":"Alex Wang, Yada Pruksachatkun, Nikita Nangia, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R. Bowman. 2019a. SuperGLUE: A stickier benchmark for general-purpose language understanding systems. In Advances in Neural Information Processing Systems 32 (NeurIPS 2019).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_4_171_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2017.0735"},{"key":"e_1_3_4_172_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chb.2019.03.014"},{"key":"e_1_3_4_173_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2019.1705513"},{"key":"e_1_3_4_174_1","volume-title":"Proceedings of the International Conference on Learning Representations 2021.","author":"Wang Zirui","year":"2021","unstructured":"Zirui Wang, Yulia Tsvetkov, Orhan Firat, and Yuan Cao. 2021a. Gradient vaccine: Investigating and improving multi-task optimization in massively multilingual models. In Proceedings of the International Conference on Learning Representations 2021."},{"key":"e_1_3_4_175_1","unstructured":"Zirui Wang Adams Wei Yu Orhan Firat and Yuan Cao. 2021b. Towards zero-label language learning. arXiv:2109.09193. Retrieved from https:\/\/arxiv.org\/abs\/2109.09193."},{"key":"e_1_3_4_176_1","unstructured":"Jason Wei Maarten Bosma Vincent Y. Zhao Kelvin Guu Adams Wei Yu Brian Lester Nan Du Andrew M. Dai and Quoc V. Le. 2021. Finetuned language models are zero-shot learners. arXiv:2109.01652 . Retrieved from https:\/\/arxiv.org\/abs\/2109.01652."},{"key":"e_1_3_4_177_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"e_1_3_4_178_1","first-page":"917","volume-title":"Proceedings of Empirical Methods in Natural Language Processing 2020","author":"Wu Chien-Sheng","year":"2020","unstructured":"Chien-Sheng Wu, Steven Hoi, Richard Socher, and Caiming Xiong. 2020. TOD-BERT: Pre-trained natural language understanding for task-oriented dialogues. In Proceedings of Empirical Methods in Natural Language Processing 2020. ACL, 917\u2013929."},{"key":"e_1_3_4_179_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2019\/14420"},{"key":"e_1_3_4_180_1","unstructured":"Jeff Wu Long Ouyang Daniel M. Ziegler Nisan Stiennon Ryan Lowe Jan Leike and Paul Christiano. 2021. Recursively summarizing books with human feedback. arXiv:2109.10862 . Retrieved from https:\/\/arxiv.org\/abs\/2109.10862."},{"key":"e_1_3_4_181_1","first-page":"7516","volume-title":"Proceedings of Empirical Methods in Natural Language Processing 2020","author":"Xia Patrick","year":"2020","unstructured":"Patrick Xia, Shijie Wu, and Benjamin Van Durme. 2020. Which* BERT? A survey organizing contextualized encoders. In Proceedings of Empirical Methods in Natural Language Processing 2020. ACL, 7516\u20137533."},{"key":"e_1_3_4_182_1","first-page":"2324","volume-title":"Proceedings of the Annual Meeting of the North American Chapter of the Association of Computational Linguistics 2019","author":"Xu Hu","year":"2019","unstructured":"Hu Xu, Bing Liu, Lei Shu, and Philip S. Yu. 2019. BERT post-training for review reading comprehension and aspect-based sentiment analysis. In Proceedings of the Annual Meeting of the North American Chapter of the Association of Computational Linguistics 2019. ACL, 2324\u20132335."},{"key":"e_1_3_4_183_1","first-page":"2048","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Xu Kelvin","year":"2015","unstructured":"Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, Ruslan Salakhudinov, Rich Zemel, and Yoshua Bengio. 2015. Show, attend and tell: Neural image caption generation with visual attention. In Proceedings of the International Conference on Machine Learning. 2048\u20132057."},{"key":"e_1_3_4_184_1","first-page":"3632","volume-title":"Proceedings of Empirical Methods in Natural Language Processing 2020","author":"Xu Yumo","year":"2020","unstructured":"Yumo Xu and Mirella Lapata. 2020. Coarse-to-fine query focused multi-document summarization. In Proceedings of Empirical Methods in Natural Language Processing 2020. ACL, 3632\u20133645."},{"key":"e_1_3_4_185_1","doi-asserted-by":"publisher","DOI":"10.1287\/isre.2017.0727"},{"key":"e_1_3_4_186_1","first-page":"1560","volume-title":"Proceedings Empirical Methods in Natural Language Processing 2020: Findings","author":"Yang Ruosong","year":"2020","unstructured":"Ruosong Yang, Jiannong Cao, Zhiyuan Wen, Youzheng Wu, and Xiaodong He. 2020. Enhancing automated essay scoring performance via cohesion measurement and combination of regression and ranking. In Proceedings Empirical Methods in Natural Language Processing 2020: Findings. ACL, 1560\u20131569."},{"key":"e_1_3_4_187_1","article-title":"XLNet: Generalized autoregressive pretraining for language understanding","volume":"32","author":"Yang Zhilin","year":"2019","unstructured":"Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Russ R. Salakhutdinov, and Quoc V. Le. 2019. XLNet: Generalized autoregressive pretraining for language understanding. In Advances in Neural Information Processing Systems 32 (NuerIPS 2019).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_4_188_1","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocz164"},{"key":"e_1_3_4_189_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2019.1661096"},{"key":"e_1_3_4_190_1","unstructured":"Holly Young. 2015. The digital language divide. The Guardian . Retrieved from http:\/\/labs.theguardian.com\/digital-language-divide\/."},{"key":"e_1_3_4_191_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2019\/13042"},{"key":"e_1_3_4_192_1","article-title":"Big Bird: Transformers for longer sequences","volume":"33","author":"Zaheer Manzil","year":"2020","unstructured":"Manzil Zaheer, Guru Guruganesh, Avinava Dubey, Joshua Ainslie, Chris Alberti, Santiago Ontanon, Philip Pham, Anirudh Ravula, Qifan Wang, Li Yang, and Amr Ahmed. 2020. Big Bird: Transformers for longer sequences. In Advances in Neural Information Processing Systems 33 (NeurIPS 2020).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_4_193_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00688"},{"key":"e_1_3_4_194_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2016.1205907"},{"key":"e_1_3_4_195_1","doi-asserted-by":"publisher","DOI":"10.5555\/3177649.3177653"},{"key":"e_1_3_4_196_1","first-page":"9563","volume-title":"Proceedings of Thirty-Fourth AAAI Conference on AI (AAAI 2020)","author":"Zhang Shuailiang","year":"2020","unstructured":"Shuailiang Zhang, Hai Zhao, Yuwei Wu, Zhuosheng Zhang, Xi Zhou, and Xiang Zhou. 2020. DCMN+: Dual co-matching network for multi-choice reading comprehension. In Proceedings of Thirty-Fourth AAAI Conference on AI (AAAI 2020). AAAI, 9563\u20139570."},{"key":"e_1_3_4_197_1","doi-asserted-by":"publisher","DOI":"10.25300\/MISQ\/2020\/15106"},{"key":"e_1_3_4_198_1","first-page":"165","volume-title":"Proceedings of Empirical Methods in Natural Language Processing 2019","author":"Zhong Peixiang","year":"2019","unstructured":"Peixiang Zhong, Di Wang, and Chunyan Miao. 2019. Knowledge-enriched transformer for emotion detection in textual conversations. In Proceedings of Empirical Methods in Natural Language Processing 2019. ACL, 165\u2013176."},{"key":"e_1_3_4_199_1","doi-asserted-by":"publisher","DOI":"10.1080\/07421222.2018.1451956"},{"key":"e_1_3_4_200_1","first-page":"3054","volume-title":"Proceedings of Empirical Methods in Natural Language Processing","author":"Zhu Junnan","year":"2019","unstructured":"Junnan Zhu, Qian Wang, Yining Wang, Yu Zhou, Jiajun Zhang, Shaonan Wang, and Chengqing Zong. 2019. NCLS: Neural cross-lingual summarization. In Proceedings of Empirical Methods in Natural Language Processing 2019. ACL 3054\u20133064."}],"container-title":["ACM Computing Surveys"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3505245","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3505245","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:31:26Z","timestamp":1750188686000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3505245"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,1,31]]},"references-count":199,"journal-issue":{"issue":"10s","published-print":{"date-parts":[[2022,1,31]]}},"alternative-id":["10.1145\/3505245"],"URL":"https:\/\/doi.org\/10.1145\/3505245","relation":{},"ISSN":["0360-0300","1557-7341"],"issn-type":[{"value":"0360-0300","type":"print"},{"value":"1557-7341","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,1,31]]},"assertion":[{"value":"2021-03-15","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2021-12-07","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2022-09-13","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}