{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T21:41:11Z","timestamp":1775252471821,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":197,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T00:00:00Z","timestamp":1655683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,21]]},"DOI":"10.1145\/3531146.3533088","type":"proceedings-article","created":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T14:27:10Z","timestamp":1655735230000},"page":"214-229","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":452,"title":["Taxonomy of Risks posed by Language Models"],"prefix":"10.1145","author":[{"given":"Laura","family":"Weidinger","sequence":"first","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Jonathan","family":"Uesato","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Maribeth","family":"Rauh","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Conor","family":"Griffin","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Po-Sen","family":"Huang","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"John","family":"Mellor","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Amelia","family":"Glaese","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Myra","family":"Cheng","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom and California Institute of Technology, USA"}]},{"given":"Borja","family":"Balle","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Atoosa","family":"Kasirzadeh","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom and University of Toronto, Canada"}]},{"given":"Courtney","family":"Biles","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Sasha","family":"Brown","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Zac","family":"Kenton","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Will","family":"Hawkins","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Tom","family":"Stepleton","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Abeba","family":"Birhane","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom and University College Dublin, Ireland"}]},{"given":"Lisa Anne","family":"Hendricks","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Laura","family":"Rimell","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"William","family":"Isaac","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Julia","family":"Haas","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Sean","family":"Legassick","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Geoffrey","family":"Irving","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]},{"given":"Iason","family":"Gabriel","sequence":"additional","affiliation":[{"name":"DeepMind, United Kingdom"}]}],"member":"320","published-online":{"date-parts":[[2022,6,20]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/2976749.2978318"},{"key":"e_1_3_2_1_2_1","volume-title":"Persistent Anti-Muslim Bias in Large Language Models. arXiv:2101.05783 [cs] (January","author":"Abid Abubakar","year":"2021","unstructured":"Abubakar Abid, Maheen Farooqi, and James Zou. 2021. Persistent Anti-Muslim Bias in Large Language Models. arXiv:2101.05783 [cs] (January 2021). http:\/\/arxiv.org\/abs\/2101.05783 arXiv:2101.05783."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.3386\/w24196"},{"key":"e_1_3_2_1_4_1","volume-title":"MasakhaNER: Named Entity Recognition for African Languages. arXiv:2103.11811 [cs] (July","author":"Adelani David\u00a0Ifeoluwa","year":"2021","unstructured":"David\u00a0Ifeoluwa Adelani, Jade Abbott, Graham Neubig, Daniel D\u2019souza, Julia Kreutzer, Constantine Lignos, Chester Palen-Michel, Happy Buzaaba, Shruti Rijhwani, Sebastian Ruder, Stephen Mayhew, Israel\u00a0Abebe Azime, Shamsuddeen Muhammad, Chris\u00a0Chinenye Emezue, Joyce Nakatumba-Nabende, Perez Ogayo, Anuoluwapo Aremu, Catherine Gitau, Derguene Mbaye, Jesujoba Alabi, Seid\u00a0Muhie Yimam, Tajuddeen Gwadabe, Ignatius Ezeani, Rubungo\u00a0Andre Niyongabo, Jonathan Mukiibi, Verrah Otiende, Iroro Orife, Davis David, Samba Ngom, Tosin Adewumi, Paul Rayson, Mofetoluwa Adeyemi, Gerald Muriuki, Emmanuel Anebi, Chiamaka Chukwuneke, Nkiruka Odu, Eric\u00a0Peter Wairagala, Samuel Oyerinde, Clemencia Siro, Tobius\u00a0Saul Bateesa, Temilola Oloyede, Yvonne Wambui, Victor Akinode, Deborah Nabagereka, Maurice Katusiime, Ayodele Awokoya, Mouhamadane MBOUP, Dibora Gebreyohannes, Henok Tilaye, Kelechi Nwaike, Degaga Wolde, Abdoulaye Faye, Blessing Sibanda, Orevaoghene Ahia, Bonaventure F.\u00a0P. Dossou, Kelechi Ogueji, Thierno\u00a0Ibrahima DIOP, Abdoulaye Diallo, Adewale Akinfaderin, Tendai Marengereke, and Salomey Osei. 2021. MasakhaNER: Named Entity Recognition for African Languages. arXiv:2103.11811 [cs] (July 2021). http:\/\/arxiv.org\/abs\/2103.11811 arXiv:2103.11811."},{"key":"e_1_3_2_1_5_1","volume-title":"Towards a Human-like Open-Domain Chatbot. arXiv:2001.09977 [cs, stat] (Feb","author":"Adiwardana Daniel","year":"2020","unstructured":"Daniel Adiwardana, Minh-Thang Luong, David\u00a0R. So, Jamie Hall, Noah Fiedel, Romal Thoppilan, Zi Yang, Apoorv Kulshreshtha, Gaurav Nemade, Yifeng Lu, and Quoc\u00a0V. Le. 2020. Towards a Human-like Open-Domain Chatbot. arXiv:2001.09977 [cs, stat] (Feb. 2020). http:\/\/arxiv.org\/abs\/2001.09977 arXiv:2001.09977."},{"key":"e_1_3_2_1_6_1","unstructured":"Blaise Ag\u00fcera\u00a0y Arcas Margaret Mitchell and Alexander Todorov. 2017. Physiognomy\u2019s New Clothes. https:\/\/medium.com\/@blaisea\/physiognomys-new-clothes-f2d4b59fdd6a"},{"key":"e_1_3_2_1_7_1","volume-title":"The Panopticon Is Already Here. The Atlantic (July","author":"Andersen Ross","year":"2020","unstructured":"Ross Andersen. 2020. The Panopticon Is Already Here. The Atlantic (July 2020). https:\/\/www.theatlantic.com\/magazine\/archive\/2020\/09\/china-ai-surveillance\/614197\/"},{"key":"e_1_3_2_1_8_1","volume-title":"CALM: Continuous Adaptive Learning for Language Modeling. arXiv:2004.03794 [cs] (April","author":"Arumae Kristjan","year":"2020","unstructured":"Kristjan Arumae and Parminder Bhatia. 2020. CALM: Continuous Adaptive Learning for Language Modeling. arXiv:2004.03794 [cs] (April 2020). http:\/\/arxiv.org\/abs\/2004.03794 arXiv:2004.03794."},{"key":"e_1_3_2_1_9_1","volume-title":"A General Language Assistant as a Laboratory for Alignment. arXiv:2112.00861 [cs] (Dec","author":"Askell Amanda","year":"2021","unstructured":"Amanda Askell, Yuntao Bai, Anna Chen, Dawn Drain, Deep Ganguli, Tom Henighan, Andy Jones, Nicholas Joseph, Ben Mann, Nova DasSarma, Nelson Elhage, Zac Hatfield-Dodds, Danny Hernandez, Jackson Kernion, Kamal Ndousse, Catherine Olsson, Dario Amodei, Tom Brown, Jack Clark, Sam McCandlish, Chris Olah, and Jared Kaplan. 2021. A General Language Assistant as a Laboratory for Alignment. arXiv:2112.00861 [cs] (Dec. 2021). http:\/\/arxiv.org\/abs\/2112.00861 arXiv:2112.00861."},{"key":"e_1_3_2_1_10_1","volume-title":"New Frontiers: The Evolving Content and Geography of New Work in the 20th Century - David Autor.","author":"Autor David","year":"2019","unstructured":"David Autor and Anna Salomons. 2019. New Frontiers: The Evolving Content and Geography of New Work in the 20th Century - David Autor. (2019). https:\/\/app.scholarsite.io\/david-autor\/articles\/new-frontiers-the-evolving-content-and-geography-of-new-work-in-the-20th-century Working Paper."},{"key":"e_1_3_2_1_11_1","volume-title":"Spinning Language Models for Propaganda-As-A-Service. arXiv:2112.05224 [cs] (Dec","author":"Bagdasaryan Eugene","year":"2021","unstructured":"Eugene Bagdasaryan and Vitaly Shmatikov. 2021. Spinning Language Models for Propaganda-As-A-Service. arXiv:2112.05224 [cs] (Dec. 2021). http:\/\/arxiv.org\/abs\/2112.05224 arXiv:2112.05224."},{"key":"e_1_3_2_1_12_1","unstructured":"Solon Barocas Moritz Hardt and Arvind Narayanan. 2019. Fairness and machine learning. fairmlbook.org. https:\/\/fairmlbook.org\/"},{"key":"e_1_3_2_1_13_1","first-page":"671","article-title":"Big Data\u2019s Disparate Impact","volume":"104","author":"Barocas Solon","year":"2016","unstructured":"Solon Barocas and Andrew\u00a0D. Selbst. 2016. Big Data\u2019s Disparate Impact. California Law Review 104 (2016), 671. https:\/\/heinonline.org\/HOL\/Page?handle=hein.journals\/calr104&id=695&div=&collection=","journal-title":"California Law Review"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.33011\/lilt.v6i.1239"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445922"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","unstructured":"Yoshua Bengio. 2008. Neural net language models. 3881\u00a0pages. http:\/\/www.scholarpedia.org\/article\/Neural_net_language_models","DOI":"10.4249\/scholarpedia.3881"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1093\/sf"},{"key":"e_1_3_2_1_19_1","first-page":"95","article-title":"I\u2019d Blush if I Could\u2019: Digital Assistants, Disembodied Cyborgs and the Problem of Gender. Word and Text","author":"Bergen Hilary","year":"2016","unstructured":"Hilary Bergen. 2016. \u2018I\u2019d Blush if I Could\u2019: Digital Assistants, Disembodied Cyborgs and the Problem of Gender. Word and Text, A Journal of Literary Studies and Linguistics VI, 01(2016), 95\u2013113. https:\/\/www.ceeol.com\/search\/article-detail?id=469884","journal-title":"A Journal of Literary Studies and Linguistics"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.findings-acl.340"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.2196\/11510"},{"key":"e_1_3_2_1_22_1","volume-title":"NLP. arXiv:2005.14050 [cs] (May","author":"Blodgett Su\u00a0Lin","year":"2020","unstructured":"Su\u00a0Lin Blodgett, Solon Barocas, Hal Daum\u00e9\u00a0III, and Hanna Wallach. 2020. Language (Technology) is Power: A Critical Survey of \u201dBias\u201d in NLP. arXiv:2005.14050 [cs] (May 2020). http:\/\/arxiv.org\/abs\/2005.14050 arXiv:2005.14050."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_24_1","volume-title":"On the Opportunities and Risks of Foundation Models. arXiv:2108.07258 [cs] (August","author":"Bommasani Rishi","year":"2021","unstructured":"Rishi Bommasani, Drew\u00a0A. Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael\u00a0S. Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, Erik Brynjolfsson, Shyamal Buch, Dallas Card, Rodrigo Castellon, Niladri Chatterji, Annie Chen, Kathleen Creel, Jared\u00a0Quincy Davis, Dora Demszky, Chris Donahue, Moussa Doumbouya, Esin Durmus, Stefano Ermon, John Etchemendy, Kawin Ethayarajh, Li Fei-Fei, Chelsea Finn, Trevor Gale, Lauren Gillespie, Karan Goel, Noah Goodman, Shelby Grossman, Neel Guha, Tatsunori Hashimoto, Peter Henderson, John Hewitt, Daniel\u00a0E. Ho, Jenny Hong, Kyle Hsu, Jing Huang, Thomas Icard, Saahil Jain, Dan Jurafsky, Pratyusha Kalluri, Siddharth Karamcheti, Geoff Keeling, Fereshte Khani, Omar Khattab, Pang\u00a0Wei Koh, Mark Krass, Ranjay Krishna, Rohith Kuditipudi, Ananya Kumar, Faisal Ladhak, Mina Lee, Tony Lee, Jure Leskovec, Isabelle Levent, Xiang\u00a0Lisa Li, Xuechen Li, Tengyu Ma, Ali Malik, Christopher\u00a0D. Manning, Suvir Mirchandani, Eric Mitchell, Zanele Munyikwa, Suraj Nair, Avanika Narayan, Deepak Narayanan, Ben Newman, Allen Nie, Juan\u00a0Carlos Niebles, Hamed Nilforoshan, Julian Nyarko, Giray Ogut, Laurel Orr, Isabel Papadimitriou, Joon\u00a0Sung Park, Chris Piech, Eva Portelance, Christopher Potts, Aditi Raghunathan, Rob Reich, Hongyu Ren, Frieda Rong, Yusuf Roohani, Camilo Ruiz, Jack Ryan, Christopher R\u00e9, Dorsa Sadigh, Shiori Sagawa, Keshav Santhanam, Andy Shih, Krishnan Srinivasan, Alex Tamkin, Rohan Taori, Armin\u00a0W. Thomas, Florian Tram\u00e8r, Rose\u00a0E. Wang, William Wang, Bohan Wu, Jiajun Wu, Yuhuai Wu, Sang\u00a0Michael Xie, Michihiro Yasunaga, Jiaxuan You, Matei Zaharia, Michael Zhang, Tianyi Zhang, Xikun Zhang, Yuhui Zhang, Lucia Zheng, Kaitlyn Zhou, and Percy Liang. 2021. On the Opportunities and Risks of Foundation Models. arXiv:2108.07258 [cs] (August 2021). http:\/\/arxiv.org\/abs\/2108.07258 arXiv:2108.07258."},{"key":"e_1_3_2_1_25_1","unstructured":"Sebastian Borgeaud Arthur Mensch Jordan Hoffmann Trevor Cai Eliza Rutherford Katie Millican George van\u00a0den Driessche Jean-Baptiste Lespiau Bogdan Damoc Aidan Clark Diego de\u00a0Las Casas Aurelia Guy Jacob Menick Roman Ring Tom Hennigan Saffron Huang Loren Maggiore Chris Jones Albin Cassirer Andy Brock Michela Paganini Geoffrey Irving Oriol Vinyals Simon Osindero Karen Simonyan Jack\u00a0W. Rae Erich Elsen and Laurent Sifre. 2022. Improving language models by retrieving from trillions of tokens. arXiv:2112.04426 [cs] (Jan. 2022). http:\/\/arxiv.org\/abs\/2112.04426 arXiv:2112.04426."},{"key":"e_1_3_2_1_26_1","volume-title":"Superintelligence: paths, dangers, strategies","author":"Bostrom Nick","unstructured":"Nick Bostrom. 2014. Superintelligence: paths, dangers, strategies. Oxford University Press, Oxford. OCLC: ocn881706835."},{"key":"e_1_3_2_1_27_1","unstructured":"Nick Bostrom 2011. Information hazards: A typology of potential harms from knowledge. Review of Contemporary Philosophy(2011) 44\u201379."},{"key":"e_1_3_2_1_28_1","volume-title":"Bowker and Susan\u00a0Leigh Star","author":"C.","year":"1999","unstructured":"Geoffrey\u00a0C. Bowker and Susan\u00a0Leigh Star. 1999. Sorting Things Out: Classification and Its Consequences. MIT Press, Cambridge, MA, USA."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1177\/105971230000800104"},{"key":"e_1_3_2_1_30_1","volume-title":"Language Models are Few-Shot Learners. arXiv:2005.14165 [cs] (July","author":"Brown B.","year":"2020","unstructured":"Tom\u00a0B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel\u00a0M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language Models are Few-Shot Learners. arXiv:2005.14165 [cs] (July 2020). http:\/\/arxiv.org\/abs\/2005.14165 arXiv:2005.14165."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1126\/science.aal4230"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_34_1","volume-title":"Extracting Training Data from Large Language Models. arXiv:2012.07805 [cs] (June","author":"Carlini Nicholas","year":"2021","unstructured":"Nicholas Carlini, Florian Tramer, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom Brown, Dawn Song, Ulfar Erlingsson, Alina Oprea, and Colin Raffel. 2021. Extracting Training Data from Large Language Models. arXiv:2012.07805 [cs] (June 2021). http:\/\/arxiv.org\/abs\/2012.07805 arXiv:2012.07805."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1007\/s13347-020-00415-6"},{"key":"e_1_3_2_1_36_1","volume-title":"Proceedings of the Second Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics, Barcelona, Spain (Online), 72\u201378","author":"Cercas\u00a0Curry Amanda","year":"2020","unstructured":"Amanda Cercas\u00a0Curry, Judy Robertson, and Verena Rieser. 2020. Conversational Assistants and Gender Stereotypes: Public Perceptions and Desiderata for Voice Personas. In Proceedings of the Second Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics, Barcelona, Spain (Online), 72\u201378. https:\/\/aclanthology.org\/2020.gebnlp-1.7"},{"key":"e_1_3_2_1_37_1","unstructured":"Mark Chen Jerry Tworek Heewoo Jun Qiming Yuan Henrique Ponde de\u00a0Oliveira Pinto Jared Kaplan Harri Edwards Yuri Burda Nicholas Joseph Greg Brockman Alex Ray Raul Puri Gretchen Krueger Michael Petrov Heidy Khlaaf Girish Sastry Pamela Mishkin Brooke Chan Scott Gray Nick Ryder Mikhail Pavlov Alethea Power Lukasz Kaiser Mohammad Bavarian Clemens Winter Philippe Tillet Felipe\u00a0Petroski Such Dave Cummings Matthias Plappert Fotios Chantzis Elizabeth Barnes Ariel Herbert-Voss William\u00a0Hebgen Guss Alex Nichol Alex Paino Nikolas Tezak Jie Tang Igor Babuschkin Suchir Balaji Shantanu Jain William Saunders Christopher Hesse Andrew\u00a0N. Carr Jan Leike Josh Achiam Vedant Misra Evan Morikawa Alec Radford Matthew Knight Miles Brundage Mira Murati Katie Mayer Peter Welinder Bob McGrew Dario Amodei Sam McCandlish Ilya Sutskever and Wojciech Zaremba. 2021. Evaluating Large Language Models Trained on Code. arXiv:2107.03374 [cs] (July 2021). http:\/\/arxiv.org\/abs\/2107.03374 arXiv:2107.03374."},{"key":"e_1_3_2_1_38_1","volume-title":"Palm: Scaling language modeling with pathways. arXiv preprint arXiv:2204.02311(2022).","author":"Chowdhery Aakanksha","year":"2022","unstructured":"Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, Hyung\u00a0Won Chung, Charles Sutton, Sebastian Gehrmann, 2022. Palm: Scaling language modeling with pathways. arXiv preprint arXiv:2204.02311(2022)."},{"key":"e_1_3_2_1_39_1","volume-title":"Atlas of AI","author":"Crawford Kate","year":"2095","unstructured":"Kate Crawford. 2021. Atlas of AI. Yale University Press. https:\/\/yalebooks.yale.edu\/book\/9780300209570\/atlas-ai"},{"key":"e_1_3_2_1_40_1","volume-title":"On Intersectionality: Essential Writings","author":"Crenshaw Kimberl\u00e9","year":"2017","unstructured":"Kimberl\u00e9 Crenshaw. 2017. On Intersectionality: Essential Writings. Books (March 2017). https:\/\/scholarship.law.columbia.edu\/books\/255"},{"key":"e_1_3_2_1_42_1","volume-title":"Plug and Play Language Models: A Simple Approach to Controlled Text Generation. arXiv:1912.02164 [cs] (March","author":"Dathathri Sumanth","year":"2020","unstructured":"Sumanth Dathathri, Andrea Madotto, Janice Lan, Jane Hung, Eric Frank, Piero Molino, Jason Yosinski, and Rosanne Liu. 2020. Plug and Play Language Models: A Simple Approach to Controlled Text Generation. arXiv:1912.02164 [cs] (March 2020). http:\/\/arxiv.org\/abs\/1912.02164 arXiv:1912.02164."},{"key":"e_1_3_2_1_43_1","volume-title":"Advances in Neural Information Processing Systems, Vol.\u00a032. Curran Associates","author":"Masson\u00a0d\u2019 Autume Cyprien","year":"2019","unstructured":"Cyprien de\u00a0Masson\u00a0d\u2019 Autume, Sebastian Ruder, Lingpeng Kong, and Dani Yogatama. 2019. Episodic Memory in Lifelong Language Learning. In Advances in Neural Information Processing Systems, Vol.\u00a032. Curran Associates, Inc.https:\/\/papers.nips.cc\/paper\/2019\/hash\/f8d2e80c1458ea2501f98a2cafadb397-Abstract.html"},{"key":"e_1_3_2_1_44_1","volume-title":"Creating Multimodal Interactive Agents with Imitation and Self-Supervised Learning. arXiv:2112.03763 [cs] (Dec","author":"Interactive Agents Team DeepMind","year":"2021","unstructured":"DeepMind Interactive Agents Team, Josh Abramson, Arun Ahuja, Arthur Brussee, Federico Carnevale, Mary Cassin, Felix Fischer, Petko Georgiev, Alex Goldin, Tim Harley, Felix Hill, Peter\u00a0C. Humphreys, Alden Hung, Jessica Landon, Timothy Lillicrap, Hamza Merzic, Alistair Muldal, Adam Santoro, Guy Scully, Tamara von Glehn, Greg Wayne, Nathaniel Wong, Chen Yan, and Rui Zhu. 2021. Creating Multimodal Interactive Agents with Imitation and Self-Supervised Learning. arXiv:2112.03763 [cs] (Dec. 2021). http:\/\/arxiv.org\/abs\/2112.03763 arXiv:2112.03763."},{"key":"e_1_3_2_1_45_1","volume-title":"Hilary Nicole, and Morgan\u00a0Klaus Scheuerman.","author":"Denton Emily","year":"2020","unstructured":"Emily Denton, Alex Hanna, Razvan Amironesei, Andrew Smart, Hilary Nicole, and Morgan\u00a0Klaus Scheuerman. 2020. Bringing the People Back In: Contesting Benchmark Machine Learning Datasets. arXiv:2007.07399 [cs] (July 2020). http:\/\/arxiv.org\/abs\/2007.07399 arXiv:2007.07399."},{"key":"e_1_3_2_1_46_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv:1810.04805 [cs] (May","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv:1810.04805 [cs] (May 2019). http:\/\/arxiv.org\/abs\/1810.04805 arXiv:1810.04805."},{"key":"e_1_3_2_1_48_1","volume-title":"Anticipating Safety Issues in E2E Conversational AI: Framework and Tooling. arXiv:2107.03451 [cs] (July","author":"Dinan Emily","year":"2021","unstructured":"Emily Dinan, Gavin Abercrombie, A.\u00a0Stevie Bergman, Shannon Spruit, Dirk Hovy, Y.-Lan Boureau, and Verena Rieser. 2021. Anticipating Safety Issues in E2E Conversational AI: Framework and Tooling. arXiv:2107.03451 [cs] (July 2021). http:\/\/arxiv.org\/abs\/2107.03451 arXiv:2107.03451."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/3278721.3278729"},{"key":"e_1_3_2_1_50_1","volume-title":"Documenting Large Webtext Corpora: A Case Study on the Colossal Clean Crawled Corpus. arXiv:2104.08758 [cs] (September","author":"Dodge Jesse","year":"2021","unstructured":"Jesse Dodge, Maarten Sap, Ana Marasovi\u0107, William Agnew, Gabriel Ilharco, Dirk Groeneveld, Margaret Mitchell, and Matt Gardner. 2021. Documenting Large Webtext Corpora: A Case Study on the Colossal Clean Crawled Corpus. arXiv:2104.08758 [cs] (September 2021). http:\/\/arxiv.org\/abs\/2104.08758 arXiv:2104.08758."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10676-016-9406-0"},{"key":"e_1_3_2_1_52_1","volume-title":"Truthful AI: Developing and governing AI that does not lie. arXiv:2110.06674 [cs] (Oct.","author":"Evans Owain","year":"2021","unstructured":"Owain Evans, Owen Cotton-Barratt, Lukas Finnveden, Adam Bales, Avital Balwit, Peter Wills, Luca Righetti, and William Saunders. 2021. Truthful AI: Developing and governing AI that does not lie. arXiv:2110.06674 [cs] (Oct. 2021). http:\/\/arxiv.org\/abs\/2110.06674 arXiv:2110.06674."},{"key":"e_1_3_2_1_53_1","volume-title":"AGI Safety Literature Review. arXiv:1805.01109 [cs] (May","author":"Everitt Tom","year":"2018","unstructured":"Tom Everitt, Gary Lea, and Marcus Hutter. 2018. AGI Safety Literature Review. arXiv:1805.01109 [cs] (May 2018). http:\/\/arxiv.org\/abs\/1805.01109 arXiv:1805.01109."},{"key":"e_1_3_2_1_54_1","volume-title":"Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity. arXiv:2101.03961 [cs] (January","author":"Fedus William","year":"2021","unstructured":"William Fedus, Barret Zoph, and Noam Shazeer. 2021. Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity. arXiv:2101.03961 [cs] (January 2021). http:\/\/arxiv.org\/abs\/2101.03961 arXiv:2101.03961."},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-39112-5_50"},{"key":"e_1_3_2_1_56_1","volume-title":"Fake news infiltrates financial markets. Financial Times (May","author":"Flood Chris","year":"2017","unstructured":"Chris Flood. 2017. Fake news infiltrates financial markets. Financial Times (May 2017). https:\/\/www.ft.com\/content\/a37e4874-2c2a-11e7-bc4b-5528796fe35c"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3232676"},{"key":"e_1_3_2_1_58_1","volume-title":"Discipline and punish: the birth of the prison","author":"Foucault Michel","year":"2009","unstructured":"Michel Foucault and Alan Sheridan. 2012. Discipline and punish: the birth of the prison. Vintage, New York. http:\/\/0-lib.myilibrary.com.catalogue.libraries.london.ac.uk?id=435863 OCLC: 817200914."},{"key":"e_1_3_2_1_59_1","volume-title":"The Challenge of Value Alignment: from Fairer Algorithms to AI Safety. arXiv:2101.06060 [cs] (January","author":"Gabriel Iason","year":"2021","unstructured":"Iason Gabriel and Vafa Ghazavi. 2021. The Challenge of Value Alignment: from Fairer Algorithms to AI Safety. arXiv:2101.06060 [cs] (January 2021). http:\/\/arxiv.org\/abs\/2101.06060 arXiv:2101.06060."},{"key":"e_1_3_2_1_60_1","volume-title":"Datasheets for Datasets. arXiv:1803.09010 [cs] (March","author":"Gebru Timnit","year":"2020","unstructured":"Timnit Gebru, Jamie Morgenstern, Briana Vecchione, Jennifer\u00a0Wortman Vaughan, Hanna Wallach, Hal Daum\u00e9\u00a0III, and Kate Crawford. 2020. Datasheets for Datasets. arXiv:1803.09010 [cs] (March 2020). http:\/\/arxiv.org\/abs\/1803.09010 arXiv:1803.09010."},{"key":"e_1_3_2_1_61_1","volume-title":"RealToxicityPrompts: Evaluating Neural Toxic Degeneration in Language Models. arXiv:2009.11462 [cs] (September","author":"Gehman Samuel","year":"2020","unstructured":"Samuel Gehman, Suchin Gururangan, Maarten Sap, Yejin Choi, and Noah\u00a0A. Smith. 2020. RealToxicityPrompts: Evaluating Neural Toxic Degeneration in Language Models. arXiv:2009.11462 [cs] (September 2020). http:\/\/arxiv.org\/abs\/2009.11462 arXiv:2009.11462."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-93372-6_28"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1177\/2053951719897945"},{"key":"e_1_3_2_1_65_1","volume-title":"Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass","author":"Gray Mary","year":"2019","unstructured":"Mary Gray and Siddarth Suri. 2019. Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass. Mariner Books. https:\/\/ghostwork.info\/"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1126\/scirobotics.aay7120"},{"key":"e_1_3_2_1_68_1","volume-title":"The Future of Digital Assistants Is Queer. Wired (Nov","author":"G\u00f3mez-Upegui Salom\u00e9","year":"2021","unstructured":"Salom\u00e9 G\u00f3mez-Upegui. 2021. The Future of Digital Assistants Is Queer. Wired (Nov. 2021). https:\/\/www.wired.com\/story\/digital-assistant-smart-device-gender-identity\/"},{"key":"e_1_3_2_1_69_1","volume-title":"AI-generated blog fooled tens of thousands. This is how he made it.MIT Technology Review (August","author":"Hao Karen","year":"2020","unstructured":"Karen Hao. 2020. A college kid\u2019s fake, AI-generated blog fooled tens of thousands. This is how he made it.MIT Technology Review (August 2020). https:\/\/www.technologyreview.com\/2020\/08\/14\/1006780\/ai-gpt-3-fake-blog-reached-top-of-hacker-news\/"},{"key":"e_1_3_2_1_70_1","volume-title":"The Haraway Reader","author":"Haraway Donna\u00a0Jeanne","unstructured":"Donna\u00a0Jeanne Haraway. 2004. The Haraway Reader. Psychology Press. Google-Books-ID: QxUr0gijyGoC."},{"key":"e_1_3_2_1_71_1","volume-title":"Equality of Opportunity in Supervised Learning. arXiv:1610.02413 [cs] (October","author":"Hardt Moritz","year":"2016","unstructured":"Moritz Hardt, Eric Price, and Nathan Srebro. 2016. Equality of Opportunity in Supervised Learning. arXiv:1610.02413 [cs] (October 2016). http:\/\/arxiv.org\/abs\/1610.02413 arXiv:1610.02413."},{"key":"e_1_3_2_1_72_1","volume-title":"Aligning AI With Shared Human Values. arXiv:2008.02275 [cs] (July","author":"Hendrycks Dan","year":"2021","unstructured":"Dan Hendrycks, Collin Burns, Steven Basart, Andrew Critch, Jerry Li, Dawn Song, and Jacob Steinhardt. 2021. Aligning AI With Shared Human Values. arXiv:2008.02275 [cs] (July 2021). http:\/\/arxiv.org\/abs\/2008.02275 arXiv:2008.02275."},{"key":"e_1_3_2_1_73_1","volume-title":"Scanning the horizon: a systematic literature review of methodologies. BMJ open 9, 5","author":"Hines Philip","year":"2019","unstructured":"Philip Hines, Li\u00a0Hiu Yu, Richard\u00a0H Guy, Angela Brand, and Marisa Papaluca-Amati. 2019. Scanning the horizon: a systematic literature review of methodologies. BMJ open 9, 5 (2019), e026764."},{"key":"e_1_3_2_1_75_1","volume-title":"Lisa\u00a0Anne Hendricks","author":"Hoffmann Jordan","year":"2022","unstructured":"Jordan Hoffmann, Sebastian Borgeaud, Arthur Mensch, Elena Buchatskaya, Trevor Cai, Eliza Rutherford, Diego de\u00a0Las Casas, Lisa\u00a0Anne Hendricks, Johannes Welbl, Aidan Clark, 2022. Training Compute-Optimal Large Language Models. arXiv preprint arXiv:2203.15556(2022)."},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300830"},{"key":"e_1_3_2_1_77_1","volume-title":"AI can help you write in the style of famous poets. Engadget (November","author":"Holt Kris","year":"2020","unstructured":"Kris Holt. 2020. Google\u2019s \u2019Verse by Verse\u2019 AI can help you write in the style of famous poets. Engadget (November 2020). https:\/\/www.engadget.com\/googles-ai-poetry-verse-by-verse-202105834.html"},{"key":"e_1_3_2_1_78_1","volume-title":"The Curious Case of Neural Text Degeneration. arXiv:1904.09751 [cs] (February","author":"Holtzman Ari","year":"2020","unstructured":"Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, and Yejin Choi. 2020. The Curious Case of Neural Text Degeneration. arXiv:1904.09751 [cs] (February 2020). http:\/\/arxiv.org\/abs\/1904.09751 arXiv:1904.09751."},{"key":"e_1_3_2_1_79_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_81_1","unstructured":"Kane Hsieh. 2019. Transformer Poetry. Paper Gains Publishing. https:\/\/papergains.co\/"},{"key":"e_1_3_2_1_82_1","volume-title":"Reducing Sentiment Bias in Language Models via Counterfactual Evaluation. arXiv:1911.03064 [cs] (October","author":"Huang Po-Sen","year":"2020","unstructured":"Po-Sen Huang, Huan Zhang, Ray Jiang, Robert Stanforth, Johannes Welbl, Jack Rae, Vishal Maini, Dani Yogatama, and Pushmeet Kohli. 2020. Reducing Sentiment Bias in Language Models via Counterfactual Evaluation. arXiv:1911.03064 [cs] (October 2020). http:\/\/arxiv.org\/abs\/1911.03064 arXiv:1911.03064."},{"key":"e_1_3_2_1_83_1","volume-title":"China \u2019employs 2 million to police internet\u2019. CNN (October","author":"Hunt Katie","year":"2013","unstructured":"Katie Hunt and CY Xu. 2013. China \u2019employs 2 million to police internet\u2019. CNN (October 2013). https:\/\/www.cnn.com\/2013\/10\/07\/world\/asia\/china-internet-monitors\/index.html publisher: CNN."},{"key":"e_1_3_2_1_84_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445918"},{"key":"e_1_3_2_1_85_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290607.3312915"},{"key":"e_1_3_2_1_86_1","volume-title":"How rising inequality hurts everyone, even the rich. Washington Post (February","author":"Ingraham Christopher","year":"2018","unstructured":"Christopher Ingraham. 2018. How rising inequality hurts everyone, even the rich. Washington Post (February 2018). https:\/\/www.washingtonpost.com\/news\/wonk\/wp\/2018\/02\/06\/how-rising-inequality-hurts-everyone-even-the-rich\/"},{"key":"e_1_3_2_1_87_1","volume-title":"International Workshop on Chatbot Research and Design. Springer, 34\u201348","author":"Ischen Carolin","year":"2019","unstructured":"Carolin Ischen, Theo Araujo, Hilde Voorveld, Guda van Noort, and Edith Smit. 2019. Privacy concerns in chatbot interactions. In International Workshop on Chatbot Research and Design. Springer, 34\u201348."},{"key":"e_1_3_2_1_88_1","volume-title":"Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering. arXiv:2007.01282 [cs] (Feb","author":"Izacard Gautier","year":"2021","unstructured":"Gautier Izacard and Edouard Grave. 2021. Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering. arXiv:2007.01282 [cs] (Feb. 2021). http:\/\/arxiv.org\/abs\/2007.01282 arXiv:2007.01282."},{"key":"e_1_3_2_1_89_1","volume-title":"Technical Report.","author":"James Letitia","unstructured":"Letitia James. 2021. How U.S. Companies & Partisans Hack Democracy to Undermine Your Voice. Technical Report. New York State Office of the Attorney General."},{"key":"e_1_3_2_1_90_1","unstructured":"Natasha Jaques Asma Ghandeharioun Judy\u00a0Hanwen Shen Craig Ferguson Agata Lapedriza Noah Jones Shixiang Gu and Rosalind Picard. 2019. Way off-policy batch deep reinforcement learning of implicit human preferences in dialog. arXiv preprint arXiv:1907.00456(2019)."},{"key":"e_1_3_2_1_91_1","doi-asserted-by":"publisher","DOI":"10.1057\/imfer.2013.7"},{"key":"e_1_3_2_1_92_1","doi-asserted-by":"publisher","DOI":"10.1163\/18756735-09701007"},{"key":"e_1_3_2_1_93_1","volume-title":"TinyBERT: Distilling BERT for Natural Language Understanding. arXiv:1909.10351 [cs] (Oct","author":"Jiao Xiaoqi","year":"2020","unstructured":"Xiaoqi Jiao, Yichun Yin, Lifeng Shang, Xin Jiang, Xiao Chen, Linlin Li, Fang Wang, and Qun Liu. 2020. TinyBERT: Distilling BERT for Natural Language Understanding. arXiv:1909.10351 [cs] (Oct. 2020). http:\/\/arxiv.org\/abs\/1909.10351 arXiv:1909.10351."},{"key":"e_1_3_2_1_94_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372829"},{"key":"e_1_3_2_1_95_1","volume-title":"The State and Fate of Linguistic Diversity and Inclusion in the NLP World. arXiv:2004.09095 [cs] (January","author":"Joshi Pratik","year":"2021","unstructured":"Pratik Joshi, Sebastin Santy, Amar Budhiraja, Kalika Bali, and Monojit Choudhury. 2021. The State and Fate of Linguistic Diversity and Inclusion in the NLP World. arXiv:2004.09095 [cs] (January 2021). http:\/\/arxiv.org\/abs\/2004.09095 arXiv:2004.09095."},{"key":"e_1_3_2_1_96_1","volume-title":"Aligning artificial intelligence with climate change mitigation. (Oct","author":"Kaack H","year":"2021","unstructured":"Lynn\u00a0H Kaack, Priya\u00a0L Donti, Emma Strubell, George Kamiya, Felix Creutzig, and David Rolnick. 2021. Aligning artificial intelligence with climate change mitigation. (Oct. 2021). https:\/\/hal.archives-ouvertes.fr\/hal-03368037"},{"key":"e_1_3_2_1_97_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_98_1","volume-title":"But Cannot Fly. arXiv:1911.03343 [cs] (May","author":"Kassner Nora","year":"2020","unstructured":"Nora Kassner and Hinrich Sch\u00fctze. 2020. Negated and Misprimed Probes for Pretrained Language Models: Birds Can Talk, But Cannot Fly. arXiv:1911.03343 [cs] (May 2020). http:\/\/arxiv.org\/abs\/1911.03343 arXiv:1911.03343."},{"key":"e_1_3_2_1_99_1","volume-title":"Alignment of Language Agents. arXiv:2103.14659 [cs] (March","author":"Kenton Zachary","year":"2021","unstructured":"Zachary Kenton, Tom Everitt, Laura Weidinger, Iason Gabriel, Vladimir Mikulik, and Geoffrey Irving. 2021. Alignment of Language Agents. arXiv:2103.14659 [cs] (March 2021). http:\/\/arxiv.org\/abs\/2103.14659 arXiv:2103.14659."},{"key":"e_1_3_2_1_100_1","doi-asserted-by":"publisher","DOI":"10.1080\/03080188.2020.1840224"},{"key":"e_1_3_2_1_101_1","volume-title":"A Distributional Approach to Controlled Text Generation. arXiv:2012.11635 [cs] (May","author":"Khalifa Muhammad","year":"2021","unstructured":"Muhammad Khalifa, Hady Elsahar, and Marc Dymetman. 2021. A Distributional Approach to Controlled Text Generation. arXiv:2012.11635 [cs] (May 2021). http:\/\/arxiv.org\/abs\/2012.11635 arXiv:2012.11635."},{"key":"e_1_3_2_1_102_1","volume-title":"Generalization through Memorization: Nearest Neighbor Language Models. arXiv:1911.00172 [cs] (Feb","author":"Khandelwal Urvashi","year":"2020","unstructured":"Urvashi Khandelwal, Omer Levy, Dan Jurafsky, Luke Zettlemoyer, and Mike Lewis. 2020. Generalization through Memorization: Nearest Neighbor Language Models. arXiv:1911.00172 [cs] (Feb. 2020). http:\/\/arxiv.org\/abs\/1911.00172 arXiv:1911.00172."},{"key":"e_1_3_2_1_103_1","volume-title":"Intersectional Bias in Hate Speech and Abusive Language Datasets. arXiv:2005.05921 [cs] (May","author":"Kim Jae\u00a0Yeon","year":"2020","unstructured":"Jae\u00a0Yeon Kim, Carlos Ortiz, Sarah Nam, Sarah Santiago, and Vivek Datta. 2020. Intersectional Bias in Hate Speech and Abusive Language Datasets. arXiv:2005.05921 [cs] (May 2020). http:\/\/arxiv.org\/abs\/2005.05921 arXiv:2005.05921."},{"key":"e_1_3_2_1_104_1","volume-title":"Anthropomorphism of computers: Is it mindful or mindless?Computers in Human Behavior 28, 1","author":"Kim Youjeong","year":"2012","unstructured":"Youjeong Kim and S\u00a0Shyam Sundar. 2012. Anthropomorphism of computers: Is it mindful or mindless?Computers in Human Behavior 28, 1 (2012), 241\u2013250."},{"key":"e_1_3_2_1_105_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2021.102643"},{"key":"e_1_3_2_1_106_1","volume-title":"Internet-Augmented Dialogue Generation. arXiv:2107.07566 [cs] (July","author":"Komeili Mojtaba","year":"2021","unstructured":"Mojtaba Komeili, Kurt Shuster, and Jason Weston. 2021. Internet-Augmented Dialogue Generation. arXiv:2107.07566 [cs] (July 2021). http:\/\/arxiv.org\/abs\/2107.07566 arXiv:2107.07566."},{"key":"e_1_3_2_1_107_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1218772110"},{"key":"e_1_3_2_1_108_1","volume-title":"GeDi: Generative Discriminator Guided Sequence Generation. arXiv:2009.06367 [cs] (Oct","author":"Krause Ben","year":"2020","unstructured":"Ben Krause, Akhilesh\u00a0Deepak Gotmare, Bryan McCann, Nitish\u00a0Shirish Keskar, Shafiq Joty, Richard Socher, and Nazneen\u00a0Fatema Rajani. 2020. GeDi: Generative Discriminator Guided Sequence Generation. arXiv:2009.06367 [cs] (Oct. 2020). http:\/\/arxiv.org\/abs\/2009.06367 arXiv:2009.06367."},{"key":"e_1_3_2_1_109_1","volume-title":"GitHub Copilot AI Is Leaking Functional API Keys. Analytics Drift (July","author":"Kulkarni Amit","year":"2021","unstructured":"Amit Kulkarni. 2021. GitHub Copilot AI Is Leaking Functional API Keys. Analytics Drift (July 2021). https:\/\/analyticsdrift.com\/github-copilot-ai-is-leaking-functional-api-keys\/"},{"key":"e_1_3_2_1_111_1","volume-title":"How Bots Broke the FCC\u2019s Public Comment System. Wired (November","author":"Lapowsky Issie","year":"2017","unstructured":"Issie Lapowsky. 2017. How Bots Broke the FCC\u2019s Public Comment System. Wired (November 2017). https:\/\/www.wired.com\/story\/bots-broke-fcc-public-comment-system\/"},{"key":"e_1_3_2_1_112_1","volume-title":"Tomas Kocisky, Sebastian Ruder, Dani Yogatama, Kris Cao, Susannah Young, and Phil Blunsom.","author":"Lazaridou Angeliki","year":"2021","unstructured":"Angeliki Lazaridou, Adhiguna Kuncoro, Elena Gribovskaya, Devang Agrawal, Adam Liska, Tayfun Terzi, Mai Gimenez, Cyprien de\u00a0Masson d\u2019Autume, Tomas Kocisky, Sebastian Ruder, Dani Yogatama, Kris Cao, Susannah Young, and Phil Blunsom. 2021. Mind the Gap: Assessing Temporal Generalization in Neural Language Models. arXiv:2102.01951 [cs] (Oct. 2021). http:\/\/arxiv.org\/abs\/2102.01951 arXiv:2102.01951."},{"key":"e_1_3_2_1_114_1","volume-title":"Deal or No Deal? End-to-End Learning for Negotiation Dialogues. arXiv:1706.05125 [cs] (June","author":"Lewis Mike","year":"2017","unstructured":"Mike Lewis, Denis Yarats, Yann\u00a0N. Dauphin, Devi Parikh, and Dhruv Batra. 2017. Deal or No Deal? End-to-End Learning for Negotiation Dialogues. arXiv:1706.05125 [cs] (June 2017). http:\/\/arxiv.org\/abs\/1706.05125 arXiv:1706.05125."},{"key":"e_1_3_2_1_115_1","volume-title":"Question and Answer Test-Train Overlap in Open-Domain Question Answering Datasets. arXiv:2008.02637 [cs] (August","author":"Lewis Patrick","year":"2020","unstructured":"Patrick Lewis, Pontus Stenetorp, and Sebastian Riedel. 2020. Question and Answer Test-Train Overlap in Open-Domain Question Answering Datasets. arXiv:2008.02637 [cs] (August 2020). http:\/\/arxiv.org\/abs\/2008.02637 arXiv:2008.02637."},{"key":"e_1_3_2_1_116_1","volume-title":"TeraPipe: Token-Level Pipeline Parallelism for Training Large-Scale Language Models. arXiv:2102.07988 [cs] (September","author":"Li Zhuohan","year":"2021","unstructured":"Zhuohan Li, Siyuan Zhuang, Shiyuan Guo, Danyang Zhuo, Hao Zhang, Dawn Song, and Ion Stoica. 2021. TeraPipe: Token-Level Pipeline Parallelism for Training Large-Scale Language Models. arXiv:2102.07988 [cs] (September 2021). http:\/\/arxiv.org\/abs\/2102.07988 arXiv:2102.07988."},{"key":"e_1_3_2_1_117_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377325.3377488"},{"key":"e_1_3_2_1_118_1","volume-title":"TruthfulQA: Measuring How Models Mimic Human Falsehoods. arXiv:2109.07958 [cs] (September","author":"Lin Stephanie","year":"2021","unstructured":"Stephanie Lin, Jacob Hilton, and Owain Evans. 2021. TruthfulQA: Measuring How Models Mimic Human Falsehoods. arXiv:2109.07958 [cs] (September 2021). http:\/\/arxiv.org\/abs\/2109.07958 arXiv:2109.07958."},{"key":"e_1_3_2_1_119_1","unstructured":"N LSE\u00a0Blog. 2017. Doxing is a toxic practice \u2013 no matter who is targeted | Media@LSE. https:\/\/blogs.lse.ac.uk\/medialse\/2017\/08\/18\/the-dangers-of-doxing-and-the-implications-for-media-regulation\/"},{"key":"e_1_3_2_1_120_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_121_1","doi-asserted-by":"publisher","DOI":"10.1007\/s13278-014-0193-5"},{"key":"e_1_3_2_1_122_1","doi-asserted-by":"publisher","DOI":"10.1145\/2046556.2046558"},{"key":"e_1_3_2_1_123_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-021-00323-0"},{"key":"e_1_3_2_1_124_1","doi-asserted-by":"publisher","DOI":"10.7771\/1481-4374.2560"},{"key":"e_1_3_2_1_125_1","volume-title":"Andrew Smart, and William\u00a0S. Isaac","author":"Martin\u00a0Jr Donald","year":"2020","unstructured":"Donald Martin\u00a0Jr., Vinodkumar Prabhakaran, Jill Kuhlberg, Andrew Smart, and William\u00a0S. Isaac. 2020. Participatory Problem Formulation for Fairer Machine Learning Through Community Based System Dynamics. arXiv:2005.07572 [cs, stat] (May 2020). http:\/\/arxiv.org\/abs\/2005.07572 arXiv:2005.07572."},{"key":"e_1_3_2_1_126_1","volume-title":"Understanding Human Impressions of Artificial Intelligence. PsyArxiv","author":"McKee Kevin","year":"2021","unstructured":"Kevin McKee, Xuechunzi Bai, and Susan Fiske. 2021. Understanding Human Impressions of Artificial Intelligence. PsyArxiv (2021). https:\/\/psyarxiv.com\/5ursp\/"},{"key":"e_1_3_2_1_128_1","unstructured":"Jacob Menick Maja Trebacz Vladimir Mikulik John Aslanides Francis Song Martin Chadwick Mia Glaese Susannah Young Lucy Campbell-Gillingham Geoffrey Irving 2022. Teaching language models to support answers with verified quotes. arXiv preprint arXiv:2203.11147(2022)."},{"key":"e_1_3_2_1_129_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2018.07.007"},{"key":"e_1_3_2_1_130_1","doi-asserted-by":"publisher","DOI":"10.1001\/jamainternmed.2016.0400"},{"key":"e_1_3_2_1_131_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0183537"},{"key":"e_1_3_2_1_132_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41545-021-00101-w"},{"key":"e_1_3_2_1_133_1","volume-title":"StereoSet: Measuring stereotypical bias in pretrained language models. arXiv:2004.09456 [cs] (April","author":"Nadeem Moin","year":"2020","unstructured":"Moin Nadeem, Anna Bethke, and Siva Reddy. 2020. StereoSet: Measuring stereotypical bias in pretrained language models. arXiv:2004.09456 [cs] (April 2020). http:\/\/arxiv.org\/abs\/2004.09456 arXiv:2004.09456."},{"key":"e_1_3_2_1_134_1","unstructured":"Reiichiro Nakano Jacob Hilton Suchir Balaji Jeff Wu Long Ouyang Christina Kim Christopher Hesse Shantanu Jain Vineet Kosaraju William Saunders 2021. WebGPT: Browser-assisted question-answering with human feedback. arXiv preprint arXiv:2112.09332(2021)."},{"key":"e_1_3_2_1_135_1","volume-title":"Proceedings of the International AAAI Conference on Web and Social Media 7, 1(2013)","author":"Nguyen Dong","year":"2013","unstructured":"Dong Nguyen, Rilana Gravel, Dolf Trieschnigg, and Theo Meder. 2013. \u201dHow Old Do You Think I Am?\u201d A Study of Language and Age in Twitter. Proceedings of the International AAAI Conference on Web and Social Media 7, 1(2013), 439\u2013448. https:\/\/ojs.aaai.org\/index.php\/ICWSM\/article\/view\/14381"},{"key":"e_1_3_2_1_136_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_137_1","doi-asserted-by":"publisher","DOI":"10.37016\/mr-2020-024"},{"key":"e_1_3_2_1_138_1","unstructured":"Google PAIR. 2019. People + AI Guidebook. Google. https:\/\/design.google\/ai-guidebook"},{"key":"e_1_3_2_1_139_1","volume-title":"The Emotional Chatbots Are Here to Probe Our Feelings. Wired (January","author":"Pardes Arielle","year":"2018","unstructured":"Arielle Pardes. 2018. The Emotional Chatbots Are Here to Probe Our Feelings. Wired (January 2018). https:\/\/www.wired.com\/story\/replika-open-source\/"},{"key":"e_1_3_2_1_140_1","doi-asserted-by":"publisher","DOI":"10.1037\/pspp0000020"},{"key":"e_1_3_2_1_141_1","volume-title":"Carbon Emissions and Large Neural Network Training. arXiv:2104.10350 [cs] (April","author":"Patterson David","year":"2021","unstructured":"David Patterson, Joseph Gonzalez, Quoc Le, Chen Liang, Lluis-Miquel Munguia, Daniel Rothchild, David So, Maud Texier, and Jeff Dean. 2021. Carbon Emissions and Large Neural Network Training. arXiv:2104.10350 [cs] (April 2021). http:\/\/arxiv.org\/abs\/2104.10350 arXiv:2104.10350."},{"key":"e_1_3_2_1_142_1","volume-title":"Conversational Agents and Natural Language Interaction: Techniques and Effective Practices. Information Science Reference - Imprint of: IGI Publishing","author":"Perez-Marin Diana","unstructured":"Diana Perez-Marin and Ismael Pascual-Nieto. 2011. Conversational Agents and Natural Language Interaction: Techniques and Effective Practices. Information Science Reference - Imprint of: IGI Publishing, Hershey, PA."},{"key":"e_1_3_2_1_143_1","volume-title":"Prospects for Reform","author":"Persily Nathaniel","unstructured":"Nathaniel Persily and Joshua\u00a0A. Tucker. 2020. Social Media and Democracy: The State of the Field, Prospects for Reform. Cambridge University Press. Google-Books-ID: TgH3DwAAQBAJ."},{"key":"e_1_3_2_1_144_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_145_1","volume-title":"Researchers made an OpenAI GPT-3 medical chatbot as an experiment. It told a mock patient to kill themselves. The Register (October","author":"Quach Katyanna","year":"2020","unstructured":"Katyanna Quach. 2020. Researchers made an OpenAI GPT-3 medical chatbot as an experiment. It told a mock patient to kill themselves. The Register (October 2020). https:\/\/www.theregister.com\/2020\/10\/28\/gpt3_medical_chatbot_experiment\/"},{"key":"e_1_3_2_1_146_1","doi-asserted-by":"publisher","DOI":"10.1109\/PASSAT"},{"key":"e_1_3_2_1_147_1","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans and Ilya Sutskever. 2018. Improving Language Understanding by Generative Pre-Training. (2018)."},{"key":"e_1_3_2_1_148_1","unstructured":"Jack\u00a0W. Rae Sebastian Borgeaud Trevor Cai Katie Millican Jordan Hoffmann Francis Song John Aslanides Sarah Henderson Roman Ring Susannah Young Eliza Rutherford Tom Hennigan Jacob Menick Albin Cassirer Richard Powell George van\u00a0den Driessche Lisa\u00a0Anne Hendricks Maribeth Rauh Po-Sen Huang Amelia Glaese Johannes Welbl Sumanth Dathathri Saffron Huang Jonathan Uesato John Mellor Irina Higgins Antonia Creswell Nat McAleese Amy Wu Erich Elsen Siddhant Jayakumar Elena Buchatskaya David Budden Esme Sutherland Karen Simonyan Michela Paganini Laurent Sifre Lena Martens Xiang\u00a0Lorraine Li Adhiguna Kuncoro Aida Nematzadeh Elena Gribovskaya Domenic Donato Angeliki Lazaridou Arthur Mensch Jean-Baptiste Lespiau Maria Tsimpoukelli Nikolai Grigorev Doug Fritz Thibault Sottiaux Mantas Pajarskas Toby Pohlen Zhitao Gong Daniel Toyama Cyprien de\u00a0Masson d\u2019Autume Yujia Li Tayfun Terzi Vladimir Mikulik Igor Babuschkin Aidan Clark Diego de\u00a0Las Casas Aurelia Guy Chris Jones James Bradbury Matthew Johnson Blake Hechtman Laura Weidinger Iason Gabriel William Isaac Ed Lockhart Simon Osindero Laura Rimell Chris Dyer Oriol Vinyals Kareem Ayoub Jeff Stanway Lorrayne Bennett Demis Hassabis Koray Kavukcuoglu and Geoffrey Irving. 2021. Scaling Language Models: Methods Analysis & Insights from Training Gopher. arXiv:2112.11446 [cs] (Dec. 2021). http:\/\/arxiv.org\/abs\/2112.11446 arXiv:2112.11446."},{"key":"e_1_3_2_1_149_1","volume-title":"Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. arXiv:1910.10683 [cs, stat] (July","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter\u00a0J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. arXiv:1910.10683 [cs, stat] (July 2020). http:\/\/arxiv.org\/abs\/1910.10683 arXiv:1910.10683."},{"key":"e_1_3_2_1_150_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patter.2020.100150"},{"key":"e_1_3_2_1_151_1","unstructured":"Inioluwa\u00a0Deborah Raji Emily\u00a0M Bender Amandalynne Paullada Emily Denton and Alex Hanna. 2021. AI and the everything in the whole wide world benchmark. arXiv preprint arXiv:2111.15366(2021)."},{"key":"e_1_3_2_1_152_1","volume-title":"Rebecca\u00a0N. White, Margaret Mitchell, Timnit Gebru, Ben Hutchinson, Jamila Smith-Loud, Daniel Theron, and Parker Barnes.","author":"Raji Inioluwa\u00a0Deborah","year":"2020","unstructured":"Inioluwa\u00a0Deborah Raji, Andrew Smart, Rebecca\u00a0N. White, Margaret Mitchell, Timnit Gebru, Ben Hutchinson, Jamila Smith-Loud, Daniel Theron, and Parker Barnes. 2020. Closing the AI Accountability Gap: Defining an End-to-End Framework for Internal Algorithmic Auditing. arXiv:2001.00973 [cs] (January 2020). http:\/\/arxiv.org\/abs\/2001.00973 arXiv:2001.00973."},{"key":"e_1_3_2_1_153_1","volume-title":"Training Production Language Models without Memorizing User Data. arXiv:2009.10031 [cs, stat] (September","author":"Ramaswamy Swaroop","year":"2020","unstructured":"Swaroop Ramaswamy, Om Thakkar, Rajiv Mathews, Galen Andrew, H.\u00a0Brendan McMahan, and Fran\u00e7oise Beaufays. 2020. Training Production Language Models without Memorizing User Data. arXiv:2009.10031 [cs, stat] (September 2020). http:\/\/arxiv.org\/abs\/2009.10031 arXiv:2009.10031."},{"key":"e_1_3_2_1_154_1","volume-title":"Zero-Shot Text-to-Image Generation. arXiv:2102.12092 [cs] (Feb","author":"Ramesh Aditya","year":"2021","unstructured":"Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. 2021. Zero-Shot Text-to-Image Generation. arXiv:2102.12092 [cs] (Feb. 2021). http:\/\/arxiv.org\/abs\/2102.12092 arXiv:2102.12092."},{"key":"e_1_3_2_1_155_1","volume-title":"Generating Fake Cyber Threat Intelligence Using Transformer-Based Models. arXiv:2102.04351 [cs] (June","author":"Ranade Priyanka","year":"2021","unstructured":"Priyanka Ranade, Aritran Piplai, Sudip Mittal, Anupam Joshi, and Tim Finin. 2021. Generating Fake Cyber Threat Intelligence Using Transformer-Based Models. arXiv:2102.04351 [cs] (June 2021). http:\/\/arxiv.org\/abs\/2102.04351 arXiv:2102.04351."},{"key":"e_1_3_2_1_156_1","volume-title":"Reclaiming Queer: Activist & Academic Rhetorics of Resistance","author":"Rand Erin","year":"2014","unstructured":"Erin Rand. 2014. Reclaiming Queer: Activist & Academic Rhetorics of Resistance. University of Alabama Press."},{"key":"e_1_3_2_1_157_1","unstructured":"Ehud Reiter. 2020. Could NLG systems injure or even kill people?https:\/\/ehudreiter.com\/2020\/10\/20\/could-nlg-systems-injure-or-even-kill-people\/"},{"key":"e_1_3_2_1_158_1","unstructured":"Matthew Rimmer. 2013. Patent-Busting: The Public Patent Foundation Gene Patents and the Seed Wars. In The Intellectual Property and Food Project Charles Lawson and Jay Sanderson (Eds.). Routledge."},{"key":"e_1_3_2_1_159_1","doi-asserted-by":"publisher","DOI":"10.1145\/3449206"},{"key":"e_1_3_2_1_160_1","unstructured":"Corby Rosset. 2020. Turing-NLG: A 17-billion-parameter language model by Microsoft. https:\/\/www.microsoft.com\/en-us\/research\/blog\/turing-nlg-a-17-billion-parameter-language-model-by-microsoft\/"},{"key":"e_1_3_2_1_161_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-15742-5_56"},{"key":"e_1_3_2_1_162_1","unstructured":"Sebastian Ruder. 2020. Why You Should Do NLP Beyond English. https:\/\/ruder.io\/nlp-beyond-english\/"},{"key":"e_1_3_2_1_163_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445896"},{"key":"e_1_3_2_1_164_1","volume-title":"Movement Pruning: Adaptive Sparsity by Fine-Tuning. arXiv:2005.07683 [cs] (Oct.","author":"Sanh Victor","year":"2020","unstructured":"Victor Sanh, Thomas Wolf, and Alexander\u00a0M. Rush. 2020. Movement Pruning: Adaptive Sparsity by Fine-Tuning. arXiv:2005.07683 [cs] (Oct. 2020). http:\/\/arxiv.org\/abs\/2005.07683 arXiv:2005.07683."},{"key":"e_1_3_2_1_165_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1163"},{"key":"e_1_3_2_1_166_1","volume-title":"Self-Diagnosis and Self-Debiasing: A Proposal for Reducing Corpus-Based Bias in NLP. arXiv:2103.00453 [cs] (Sept","author":"Schick Timo","year":"2021","unstructured":"Timo Schick, Sahana Udupa, and Hinrich Sch\u00fctze. 2021. Self-Diagnosis and Self-Debiasing: A Proposal for Reducing Corpus-Based Bias in NLP. arXiv:2103.00453 [cs] (Sept. 2021). http:\/\/arxiv.org\/abs\/2103.00453 arXiv:2103.00453."},{"key":"e_1_3_2_1_167_1","doi-asserted-by":"publisher","DOI":"10.1145\/3381831"},{"key":"e_1_3_2_1_169_1","volume-title":"Societal Biases in Language Generation: Progress and Challenges. arXiv:2105.04054 [cs] (June","author":"Sheng Emily","year":"2021","unstructured":"Emily Sheng, Kai-Wei Chang, Premkumar Natarajan, and Nanyun Peng. 2021. Societal Biases in Language Generation: Progress and Challenges. arXiv:2105.04054 [cs] (June 2021). http:\/\/arxiv.org\/abs\/2105.04054 arXiv:2105.04054."},{"key":"e_1_3_2_1_170_1","volume-title":"Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism. arXiv:1909.08053 [cs] (March","author":"Shoeybi Mohammad","year":"2020","unstructured":"Mohammad Shoeybi, Mostofa Patwary, Raul Puri, Patrick LeGresley, Jared Casper, and Bryan Catanzaro. 2020. Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism. arXiv:1909.08053 [cs] (March 2020). http:\/\/arxiv.org\/abs\/1909.08053 arXiv:1909.08053."},{"key":"e_1_3_2_1_171_1","volume-title":"Process for Adapting Language Models to Society (PALMS) with Values-Targeted Datasets. arXiv:2106.10328 [cs] (June","author":"Solaiman Irene","year":"2021","unstructured":"Irene Solaiman and Christy Dennison. 2021. Process for Adapting Language Models to Society (PALMS) with Values-Targeted Datasets. arXiv:2106.10328 [cs] (June 2021). http:\/\/arxiv.org\/abs\/2106.10328 arXiv:2106.10328."},{"key":"e_1_3_2_1_172_1","volume-title":"Language modelling\u2019s generative model: is it rational?Computer Laboratory","author":"Sparck\u00a0Jones Karen","unstructured":"Karen Sparck\u00a0Jones. 2004. Language modelling\u2019s generative model: is it rational?Computer Laboratory, University of Cambridge, Cambridge, UK."},{"key":"e_1_3_2_1_173_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10676-016-9392-2"},{"key":"e_1_3_2_1_174_1","volume-title":"The Coal Question: An Inquiry Concerning the Progress of the Nation, and the Probable Exhaustion of Our Coal-mines(3 ed.). Augustus M","unstructured":"William. Stanley\u00a0Jevons. 1905. The Coal Question: An Inquiry Concerning the Progress of the Nation, and the Probable Exhaustion of Our Coal-mines(3 ed.). Augustus M. Kelley, New York."},{"key":"e_1_3_2_1_175_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.respol.2013.05.008"},{"key":"e_1_3_2_1_176_1","volume-title":"Energy and Policy Considerations for Deep Learning in NLP. arXiv:1906.02243 [cs] (June","author":"Strubell Emma","year":"2019","unstructured":"Emma Strubell, Ananya Ganesh, and Andrew McCallum. 2019. Energy and Policy Considerations for Deep Learning in NLP. arXiv:1906.02243 [cs] (June 2019). http:\/\/arxiv.org\/abs\/1906.02243 arXiv:1906.02243."},{"key":"e_1_3_2_1_177_1","doi-asserted-by":"crossref","unstructured":"Shannon Sullivan and Nancy Tuana (Eds.). 2007. Race and epistemologies of ignorance. State University of New York Press Albany. OCLC: ocm70676503.","DOI":"10.1353\/book5200"},{"key":"e_1_3_2_1_178_1","unstructured":"summerstay on Reddit. 2020. Fiction by Neil Gaiman and Terry Pratchett by GPT-3. www.reddit.com\/r\/slatestarcodex\/comments\/hmu5lm\/fiction_by_neil_gaiman_and_terry_pratchett_by_gpt3\/"},{"key":"e_1_3_2_1_179_1","volume-title":"LAMOL: LAnguage MOdeling for Lifelong Language Learning. arXiv:1909.03329 [cs] (Dec.","author":"Sun Fan-Keng","year":"2019","unstructured":"Fan-Keng Sun, Cheng-Hao Ho, and Hung-Yi Lee. 2019. LAMOL: LAnguage MOdeling for Lifelong Language Learning. arXiv:1909.03329 [cs] (Dec. 2019). http:\/\/arxiv.org\/abs\/1909.03329 arXiv:1909.03329."},{"key":"e_1_3_2_1_180_1","volume-title":"Understanding the Capabilities, Limitations, and Societal Impact of Large Language Models. arXiv:2102.02503 [cs] (February","author":"Tamkin Alex","year":"2021","unstructured":"Alex Tamkin, Miles Brundage, Jack Clark, and Deep Ganguli. 2021. Understanding the Capabilities, Limitations, and Societal Impact of Large Language Models. arXiv:2102.02503 [cs] (February 2021). http:\/\/arxiv.org\/abs\/2102.02503 arXiv:2102.02503."},{"key":"e_1_3_2_1_181_1","doi-asserted-by":"publisher","DOI":"10.1145\/3461702.3462540"},{"key":"e_1_3_2_1_182_1","volume-title":"Facebook AI WMT21 News Translation Task Submission. arXiv:2108.03265 [cs] (Aug","author":"Tran Chau","year":"2021","unstructured":"Chau Tran, Shruti Bhosale, James Cross, Philipp Koehn, Sergey Edunov, and Angela Fan. 2021. Facebook AI WMT21 News Translation Task Submission. arXiv:2108.03265 [cs] (Aug. 2021). http:\/\/arxiv.org\/abs\/2108.03265 arXiv:2108.03265."},{"key":"e_1_3_2_1_183_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chb.2019.04.009"},{"key":"e_1_3_2_1_184_1","unstructured":"VersebyVerse 2020. Verse by Verse. https:\/\/sites.research.google\/versebyverse\/publisher:."},{"key":"e_1_3_2_1_185_1","volume-title":"The invention of AI \u2018gaydar","author":"Vincent James","year":"2017","unstructured":"James Vincent. 2017. The invention of AI \u2018gaydar\u2019 could be the start of something much worse. The Verge (September 2017). https:\/\/www.theverge.com\/2017\/9\/21\/16332760\/ai-sexuality-gaydar-photo-physiognomy"},{"key":"e_1_3_2_1_186_1","doi-asserted-by":"publisher","DOI":"10.1111\/jopp.12262"},{"key":"e_1_3_2_1_187_1","volume-title":"Privacy matters because it empowers us all | Aeon Essays. Aeon (September","author":"V\u00e9liz Carissa","year":"2019","unstructured":"Carissa V\u00e9liz. 2019. Privacy matters because it empowers us all | Aeon Essays. Aeon (September 2019). https:\/\/aeon.co\/essays\/privacy-matters-because-it-empowers-us-all"},{"key":"e_1_3_2_1_188_1","unstructured":"Daniel Wallace Florian Tramer Matthew Jagielski and Ariel Herbert-Voss. 2020. Does GPT-2 Know Your Phone Number?http:\/\/bair.berkeley.edu\/blog\/2020\/12\/20\/lmmem\/"},{"key":"e_1_3_2_1_189_1","volume-title":"MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers. arXiv:2002.10957 [cs] (April","author":"Wang Wenhui","year":"2020","unstructured":"Wenhui Wang, Furu Wei, Li Dong, Hangbo Bao, Nan Yang, and Ming Zhou. 2020. MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers. arXiv:2002.10957 [cs] (April 2020). http:\/\/arxiv.org\/abs\/2002.10957 arXiv:2002.10957."},{"key":"e_1_3_2_1_190_1","doi-asserted-by":"publisher","DOI":"10.1037\/pspa0000098"},{"key":"e_1_3_2_1_191_1","doi-asserted-by":"publisher","DOI":"10.5555\/2390374.2390377"},{"key":"e_1_3_2_1_192_1","volume-title":"The Impact of Artificial Intelligence on the Labor Market. SSRN Scholarly Paper ID 3482150. Social Science Research Network","author":"Webb Michael","unstructured":"Michael Webb. 2019. The Impact of Artificial Intelligence on the Labor Market. SSRN Scholarly Paper ID 3482150. Social Science Research Network, Rochester, NY. https:\/\/papers.ssrn.com\/abstract=3482150"},{"key":"e_1_3_2_1_193_1","volume-title":"Ethical and social risks of harm from Language Models. arXiv:2112.04359 [cs] (Dec","author":"Weidinger Laura","year":"2021","unstructured":"Laura Weidinger, John Mellor, Maribeth Rauh, Conor Griffin, Jonathan Uesato, Po-Sen Huang, Myra Cheng, Mia Glaese, Borja Balle, Atoosa Kasirzadeh, Zac Kenton, Sasha Brown, Will Hawkins, Tom Stepleton, Courtney Biles, Abeba Birhane, Julia Haas, Laura Rimell, Lisa\u00a0Anne Hendricks, William Isaac, Sean Legassick, Geoffrey Irving, and Iason Gabriel. 2021. Ethical and social risks of harm from Language Models. arXiv:2112.04359 [cs] (Dec. 2021). http:\/\/arxiv.org\/abs\/2112.04359 arXiv:2112.04359."},{"key":"e_1_3_2_1_194_1","volume-title":"Challenges in Detoxifying Language Models. arXiv:2109.07445 [cs] (September","author":"Welbl Johannes","year":"2021","unstructured":"Johannes Welbl, Amelia Glaese, Jonathan Uesato, Sumanth Dathathri, John Mellor, Lisa\u00a0Anne Hendricks, Kirsty Anderson, Pushmeet Kohli, Ben Coppin, and Po-Sen Huang. 2021. Challenges in Detoxifying Language Models. arXiv:2109.07445 [cs] (September 2021). http:\/\/arxiv.org\/abs\/2109.07445 arXiv:2109.07445."},{"key":"e_1_3_2_1_196_1","volume-title":"Language Models are Few-shot Multilingual Learners. arXiv:2109.07684 [cs] (September","author":"Winata Genta\u00a0Indra","year":"2021","unstructured":"Genta\u00a0Indra Winata, Andrea Madotto, Zhaojiang Lin, Rosanne Liu, Jason Yosinski, and Pascale Fung. 2021. Language Models are Few-shot Multilingual Learners. arXiv:2109.07684 [cs] (September 2021). http:\/\/arxiv.org\/abs\/2109.07684 arXiv:2109.07684."},{"key":"e_1_3_2_1_197_1","volume-title":"Detoxifying Language Models Risks Marginalizing Minority Voices. arXiv:2104.06390 [cs] (April","author":"Xu Albert","year":"2021","unstructured":"Albert Xu, Eshaan Pathak, Eric Wallace, Suchin Gururangan, Maarten Sap, and Dan Klein. 2021. Detoxifying Language Models Risks Marginalizing Minority Voices. arXiv:2104.06390 [cs] (April 2021). http:\/\/arxiv.org\/abs\/2104.06390 arXiv:2104.06390."},{"key":"e_1_3_2_1_198_1","volume-title":"mT5: A massively multilingual pre-trained text-to-text transformer. arXiv:2010.11934 [cs] (March","author":"Xue Linting","year":"2021","unstructured":"Linting Xue, Noah Constant, Adam Roberts, Mihir Kale, Rami Al-Rfou, Aditya Siddhant, Aditya Barua, and Colin Raffel. 2021. mT5: A massively multilingual pre-trained text-to-text transformer. arXiv:2010.11934 [cs] (March 2021). http:\/\/arxiv.org\/abs\/2010.11934 arXiv:2010.11934."},{"key":"e_1_3_2_1_199_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.0021-8529.2005.00190.x"},{"key":"e_1_3_2_1_200_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1418680112"},{"key":"e_1_3_2_1_201_1","volume-title":"Differentially Private Fine-tuning of Language Models. arXiv:2110.06500 [cs, stat] (Oct","author":"Yu Da","year":"2021","unstructured":"Da Yu, Saurabh Naik, Arturs Backurs, Sivakanth Gopi, Huseyin\u00a0A. Inan, Gautam Kamath, Janardhan Kulkarni, Yin\u00a0Tat Lee, Andre Manoel, Lukas Wutschitz, Sergey Yekhanin, and Huishuai Zhang. 2021. Differentially Private Fine-tuning of Language Models. arXiv:2110.06500 [cs, stat] (Oct. 2021). http:\/\/arxiv.org\/abs\/2110.06500 arXiv:2110.06500."},{"key":"e_1_3_2_1_202_1","doi-asserted-by":"publisher","DOI":"10.1080\/10572250701380766"},{"key":"e_1_3_2_1_203_1","volume-title":"Defending Against Neural Fake News. arXiv:1905.12616 [cs] (Dec","author":"Zellers Rowan","year":"2020","unstructured":"Rowan Zellers, Ari Holtzman, Hannah Rashkin, Yonatan Bisk, Ali Farhadi, Franziska Roesner, and Yejin Choi. 2020. Defending Against Neural Fake News. arXiv:1905.12616 [cs] (Dec. 2020). http:\/\/arxiv.org\/abs\/1905.12616 arXiv:1905.12616."},{"key":"e_1_3_2_1_204_1","volume-title":"Differentiable Prompt Makes Pre-trained Language Models Better Few-shot Learners. arXiv:2108.13161 [cs] (October","author":"Zhang Ningyu","year":"2021","unstructured":"Ningyu Zhang, Luoqiu Li, Xiang Chen, Shumin Deng, Zhen Bi, Chuanqi Tan, Fei Huang, and Huajun Chen. 2021. Differentiable Prompt Makes Pre-trained Language Models Better Few-shot Learners. arXiv:2108.13161 [cs] (October 2021). http:\/\/arxiv.org\/abs\/2108.13161 arXiv:2108.13161."},{"key":"e_1_3_2_1_205_1","doi-asserted-by":"publisher","DOI":"10.48550\/arxiv.2205.01068"},{"key":"e_1_3_2_1_206_1","volume-title":"Calibrate Before Use: Improving Few-Shot Performance of Language Models. arXiv:2102.09690 [cs] (June","author":"Zhao Z.","year":"2021","unstructured":"Tony\u00a0Z. Zhao, Eric Wallace, Shi Feng, Dan Klein, and Sameer Singh. 2021. Calibrate Before Use: Improving Few-Shot Performance of Language Models. arXiv:2102.09690 [cs] (June 2021). http:\/\/arxiv.org\/abs\/2102.09690 arXiv:2102.09690."},{"key":"e_1_3_2_1_207_1","unstructured":"Daniel\u00a0M Ziegler Nisan Stiennon Jeffrey Wu Tom\u00a0B Brown Alec Radford Dario Amodei Paul Christiano and Geoffrey Irving. 2019. Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593(2019)."},{"key":"e_1_3_2_1_208_1","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-014-0267-6"}],"event":{"name":"FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency","location":"Seoul Republic of Korea","acronym":"FAccT '22","sponsor":["ACM Association for Computing Machinery"]},"container-title":["2022 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533088","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3533088","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:02:10Z","timestamp":1750186930000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533088"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,20]]},"references-count":197,"alternative-id":["10.1145\/3531146.3533088","10.1145\/3531146"],"URL":"https:\/\/doi.org\/10.1145\/3531146.3533088","relation":{},"subject":[],"published":{"date-parts":[[2022,6,20]]},"assertion":[{"value":"2022-06-20","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}