{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,25]],"date-time":"2026-04-25T12:11:43Z","timestamp":1777119103100,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":83,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,4,19]],"date-time":"2023-04-19T00:00:00Z","timestamp":1681862400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["318937"],"award-info":[{"award-number":["318937"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"name":"European Commission Horizon 2020 FET Proactive program","award":["101017779"],"award-info":[{"award-number":["101017779"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,4,19]]},"DOI":"10.1145\/3544548.3580688","type":"proceedings-article","created":{"date-parts":[[2023,4,20]],"date-time":"2023-04-20T04:26:08Z","timestamp":1681964768000},"page":"1-19","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":209,"title":["Evaluating Large Language Models in Generating Synthetic HCI Research Data: a Case Study"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7764-3459","authenticated-orcid":false,"given":"Perttu","family":"H\u00e4m\u00e4l\u00e4inen","sequence":"first","affiliation":[{"name":"Aalto University, Finland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0671-5755","authenticated-orcid":false,"given":"Mikke","family":"Tavast","sequence":"additional","affiliation":[{"name":"Aalto University, Finland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2951-6399","authenticated-orcid":false,"given":"Anton","family":"Kunnari","sequence":"additional","affiliation":[{"name":"University of Helsinki, Finland"}]}],"member":"320","published-online":{"date-parts":[[2023,4,19]]},"reference":[{"key":"e_1_3_3_3_1_1","doi-asserted-by":"publisher","DOI":"10.3758\/s13428-019-01237-x"},{"key":"e_1_3_3_3_2_1","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2209.06899"},{"key":"e_1_3_3_3_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/1978942.1979336"},{"key":"e_1_3_3_3_4_1","doi-asserted-by":"publisher","DOI":"10.1038\/nbt.4314"},{"key":"e_1_3_3_3_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2013.50"},{"key":"e_1_3_3_3_6_1","doi-asserted-by":"publisher","unstructured":"Julia Bopp Jan\u00a0Benjamin Vornhagen Roosa Piitulainen Barbara Keller and Elisa\u00a0D. Mekler. 2020. GamesAsArt. (July 2020). https:\/\/doi.org\/10.17605\/OSF.IO\/RYVT6 Publisher: OSF.","DOI":"10.17605\/OSF.IO"},{"key":"e_1_3_3_3_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3474664"},{"key":"e_1_3_3_3_8_1","unstructured":"Gwern Branwen. 2020. GPT-3 creative fiction. (2020). https:\/\/www.gwern.net\/GPT-3"},{"key":"e_1_3_3_3_9_1","doi-asserted-by":"publisher","DOI":"10.1191\/1478088706qp063oa"},{"key":"e_1_3_3_3_10_1","volume-title":"Advances in Neural Information Processing Systems, H.\u00a0Larochelle, M.\u00a0Ranzato, R.\u00a0Hadsell, M.\u00a0F. Balcan, and H.\u00a0Lin (Eds.). Vol.\u00a033. Curran Associates","author":"Brown Tom","year":"1877","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared\u00a0D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Chris Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language Models are Few-Shot Learners. In Advances in Neural Information Processing Systems, H.\u00a0Larochelle, M.\u00a0Ranzato, R.\u00a0Hadsell, M.\u00a0F. Balcan, and H.\u00a0Lin (Eds.). Vol.\u00a033. Curran Associates, Inc., 1877\u20131901. https:\/\/proceedings.neurips.cc\/paper\/2020\/file\/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf"},{"key":"e_1_3_3_3_11_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jml.2019.104047"},{"key":"e_1_3_3_3_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/MCI.2014.2307227"},{"key":"e_1_3_3_3_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376701"},{"key":"e_1_3_3_3_14_1","unstructured":"Aakanksha Chowdhery Sharan Narang Jacob Devlin Maarten Bosma Gaurav Mishra Adam Roberts Paul Barham Hyung\u00a0Won Chung Charles Sutton Sebastian Gehrmann Parker Schuh Kensen Shi Sasha Tsvyashchenko Joshua Maynez Abhishek Rao Parker Barnes Yi Tay Noam Shazeer Vinodkumar Prabhakaran Emily Reif Nan Du Ben Hutchinson Reiner Pope James Bradbury Jacob Austin Michael Isard Guy Gur-Ari Pengcheng Yin Toju Duke Anselm Levskaya Sanjay Ghemawat Sunipa Dev Henryk Michalewski Xavier Garcia Vedant Misra Kevin Robinson Liam Fedus Denny Zhou Daphne Ippolito David Luan Hyeontaek Lim Barret Zoph Alexander Spiridonov Ryan Sepassi David Dohan Shivani Agrawal Mark Omernick Andrew\u00a0M. Dai Thanumalayan\u00a0Sankaranarayana Pillai Marie Pellat Aitor Lewkowycz Erica Moreira Rewon Child Oleksandr Polozov Katherine Lee Zongwei Zhou Xuezhi Wang Brennan Saeta Mark Diaz Orhan Firat Michele Catasta Jason Wei Kathy Meier-Hellstern Douglas Eck Jeff Dean Slav Petrov and Noah Fiedel. 2022. PaLM: Scaling Language Modeling with Pathways. arxiv:2204.02311\u00a0[cs.CL]"},{"key":"e_1_3_3_3_15_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2020"},{"key":"e_1_3_3_3_16_1","unstructured":"Matt Cox. 2019. This AI text adventure generator lets you do anything you\u00a0want. https:\/\/www.rockpapershotgun.com\/this-ai-text-adventure-generator-lets-you-do-anything-you-want"},{"key":"e_1_3_3_3_17_1","unstructured":"Jonas Degrave. 2022. Building A Virtual Machine inside ChatGPT. https:\/\/www.engraved.blog\/building-a-virtual-machine-inside\/"},{"key":"e_1_3_3_3_18_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.naacl-main.122"},{"key":"e_1_3_3_3_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445924"},{"key":"e_1_3_3_3_20_1","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.2005.00341"},{"key":"e_1_3_3_3_21_1","unstructured":"Nelson Elhage Neel Nanda Catherine Olsson Tom Henighan Nicholas Joseph Ben Mann Amanda Askell Yuntao Bai Anna Chen Tom Conerly Nova DasSarma Dawn Drain Deep Ganguli Zac Hatfield-Dodds Danny Hernandez Andy Jones Jackson Kernion Liane Lovitt Kamal Ndousse Dario Amodei Tom Brown Jack Clark Jared Kaplan Sam McCandlish and Chris Olah. 2021. A Mathematical Framework for Transformer Circuits. Transformer Circuits Thread(2021). https:\/\/transformer-circuits.pub\/2021\/framework\/index.html."},{"key":"e_1_3_3_3_22_1","doi-asserted-by":"publisher","DOI":"10.5555\/3001460.3001507"},{"key":"e_1_3_3_3_23_1","volume-title":"The information capacity of the human motor system in controlling the amplitude of movement.Journal of experimental psychology 47, 6","author":"Fitts M","year":"1954","unstructured":"Paul\u00a0M Fitts. 1954. The information capacity of the human motor system in controlling the amplitude of movement.Journal of experimental psychology 47, 6 (1954), 381\u2013391."},{"key":"e_1_3_3_3_24_1","unstructured":"Leo Gao. 2021. On the Sizes of OpenAI API Models.Retrieved 2022-09-12 from https:\/\/blog.eleuther.ai\/gpt3-model-sizes\/"},{"key":"e_1_3_3_3_25_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11135-021-01252-1"},{"key":"e_1_3_3_3_26_1","doi-asserted-by":"publisher","unstructured":"Maarten Grootendorst. 2022. BERTopic: Neural topic modeling with a class-based TF-IDF procedure. arXiv preprint arXiv:2203.05794(2022). https:\/\/doi.org\/10.48550\/ARXIV.2203.05794","DOI":"10.48550\/ARXIV.2203.05794"},{"key":"e_1_3_3_3_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3490100.3516458"},{"key":"e_1_3_3_3_28_1","volume-title":"Advances in Neural Information Processing Systems, I.\u00a0Guyon, U.\u00a0Von Luxburg, S.\u00a0Bengio, H.\u00a0Wallach, R.\u00a0Fergus, S.\u00a0Vishwanathan, and R.\u00a0Garnett (Eds.). Vol.\u00a030. Curran Associates","author":"Heusel Martin","year":"2017","unstructured":"Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. 2017. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. In Advances in Neural Information Processing Systems, I.\u00a0Guyon, U.\u00a0Von Luxburg, S.\u00a0Bengio, H.\u00a0Wallach, R.\u00a0Fergus, S.\u00a0Vishwanathan, and R.\u00a0Garnett (Eds.). Vol.\u00a030. Curran Associates, Inc.https:\/\/proceedings.neurips.cc\/paper\/2017\/file\/8a1d694707eb0fefe65871369074926d-Paper.pdf"},{"key":"e_1_3_3_3_29_1","volume-title":"International Conference on Learning Representations.","author":"Holtzman Ari","year":"2020","unstructured":"Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, and Yejin Choi. 2020. The curious case of neural text degeneration. In International Conference on Learning Representations."},{"key":"e_1_3_3_3_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445483"},{"key":"e_1_3_3_3_31_1","doi-asserted-by":"crossref","unstructured":"Daniel Kahneman and Amos Tversky. 2013. Prospect theory: An analysis of decision under risk. In Handbook of the fundamentals of financial decision making: Part I. World Scientific 99\u2013127.","DOI":"10.1142\/9789814417358_0006"},{"key":"e_1_3_3_3_32_1","doi-asserted-by":"publisher","unstructured":"Jared Kaplan Sam McCandlish Tom Henighan Tom\u00a0B Brown Benjamin Chess Rewon Child Scott Gray Alec Radford Jeffrey Wu and Dario Amodei. 2020. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361(2020). https:\/\/doi.org\/10.48550\/ARXIV.2001.08361","DOI":"10.48550\/ARXIV.2001.08361"},{"key":"e_1_3_3_3_33_1","unstructured":"Takeshi Kojima Shixiang\u00a0Shane Gu Machel Reid Yutaka Matsuo and Yusuke Iwasawa. 2022. Large Language Models are Zero-Shot Reasoners. In Advances in Neural Information Processing Systems. https:\/\/arxiv.org\/abs\/2205.11916"},{"key":"e_1_3_3_3_34_1","volume-title":"Computational Interaction","author":"Kristensson Per\u00a0Ola","unstructured":"Per\u00a0Ola Kristensson. 2018. Statistical Language Processing for Text Entry. In Computational Interaction. Oxford University Press, 43\u201364."},{"key":"e_1_3_3_3_35_1","doi-asserted-by":"publisher","DOI":"10.1037\/xhp0000932"},{"key":"e_1_3_3_3_36_1","volume-title":"Proceedings of the Annual Meeting of the Cognitive Science Society, Vol.\u00a044","author":"Kwon Mijin","year":"2022","unstructured":"Mijin Kwon, Tor Wager, and Jonathan Phillips. 2022. Representations of emotion concepts: Comparison across pairwise, appraisal feature-based, and word embedding-based similarity spaces. In Proceedings of the Annual Meeting of the Cognitive Science Society, Vol.\u00a044."},{"key":"e_1_3_3_3_37_1","volume-title":"Advances in Neural Information Processing Systems, H.\u00a0Wallach, H.\u00a0Larochelle, A.\u00a0Beygelzimer, F.\u00a0d'Alch\u00e9-Buc, E.\u00a0Fox, and R.\u00a0Garnett (Eds.). Vol.\u00a032. Curran Associates","author":"Kynk\u00e4\u00e4nniemi Tuomas","year":"2019","unstructured":"Tuomas Kynk\u00e4\u00e4nniemi, Tero Karras, Samuli Laine, Jaakko Lehtinen, and Timo Aila. 2019. Improved Precision and Recall Metric for Assessing Generative Models. In Advances in Neural Information Processing Systems, H.\u00a0Wallach, H.\u00a0Larochelle, A.\u00a0Beygelzimer, F.\u00a0d'Alch\u00e9-Buc, E.\u00a0Fox, and R.\u00a0Garnett (Eds.). Vol.\u00a032. Curran Associates, Inc.https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/0234c510bc6d908b28c70ff313743079-Paper.pdf"},{"key":"e_1_3_3_3_38_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.isci.2021.103364"},{"key":"e_1_3_3_3_39_1","volume-title":"Deep Learning For Symbolic Mathematics. In International Conference on Learning Representations.","author":"Lample Guillaume","year":"2019","unstructured":"Guillaume Lample and Fran\u00e7ois Charton. 2019. Deep Learning For Symbolic Mathematics. In International Conference on Learning Representations."},{"key":"e_1_3_3_3_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3031549"},{"key":"e_1_3_3_3_41_1","volume-title":"International Conference on Machine Learning. PMLR, 6565\u20136576","author":"Liang Paul\u00a0Pu","year":"2021","unstructured":"Paul\u00a0Pu Liang, Chiyu Wu, Louis-Philippe Morency, and Ruslan Salakhutdinov. 2021. Towards understanding and mitigating social biases in language models. In International Conference on Machine Learning. PMLR, 6565\u20136576."},{"key":"e_1_3_3_3_42_1","volume-title":"Visual exploration of semantic relationships in neural word embeddings","author":"Liu Shusen","year":"2017","unstructured":"Shusen Liu, Peer-Timo Bremer, Jayaraman\u00a0J Thiagarajan, Vivek Srikumar, Bei Wang, Yarden Livnat, and Valerio Pascucci. 2017. Visual exploration of semantic relationships in neural word embeddings. IEEE transactions on visualization and computer graphics 24, 1(2017), 553\u2013562."},{"key":"e_1_3_3_3_43_1","unstructured":"Ziming Liu Ouail Kitouni Niklas Nolte Eric\u00a0J Michaud Max Tegmark and Mike Williams. 2022. Towards Understanding Grokking: An Effective Theory of Representation Learning. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_3_3_44_1","unstructured":"R\u00f3is\u00edn Loughran and Michael O\u2019Neill. 2017. Application Domains Considered in Computational Creativity.. In ICCC. 197\u2013204."},{"key":"e_1_3_3_3_45_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i7.20729"},{"key":"e_1_3_3_3_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/142750.142794"},{"key":"e_1_3_3_3_47_1","volume-title":"Detection theory : a user\u2019s guide","author":"Macmillan A.","unstructured":"Neil\u00a0A. Macmillan. 2005. Detection theory : a user\u2019s guide(2nd ed.). Lawrence Erlbaum Associates, Mahwah, N.J.","edition":"2"},{"key":"e_1_3_3_3_48_1","volume-title":"Progen: Language modeling for protein generation. arXiv preprint arXiv:2004.03497(2020).","author":"Madani Ali","year":"2020","unstructured":"Ali Madani, Bryan McCann, Nikhil Naik, Nitish\u00a0Shirish Keskar, Namrata Anand, Raphael\u00a0R Eguchi, Po-Ssu Huang, and Richard Socher. 2020. Progen: Language modeling for protein generation. arXiv preprint arXiv:2004.03497(2020)."},{"key":"e_1_3_3_3_49_1","doi-asserted-by":"publisher","DOI":"10.21105\/joss.00205"},{"key":"e_1_3_3_3_50_1","volume-title":"Umap: Uniform manifold approximation and projection for dimension reduction. arXiv preprint arXiv:1802.03426(2018).","author":"McInnes Leland","year":"2018","unstructured":"Leland McInnes, John Healy, and James Melville. 2018. Umap: Uniform manifold approximation and projection for dimension reduction. arXiv preprint arXiv:1802.03426(2018)."},{"key":"e_1_3_3_3_51_1","volume-title":"Proceedings of the 2013 conference of the north american chapter of the association for computational linguistics: Human language technologies. 746\u2013751","author":"Mikolov Tom\u00e1\u0161","year":"2013","unstructured":"Tom\u00e1\u0161 Mikolov, Wen-tau Yih, and Geoffrey Zweig. 2013. Linguistic regularities in continuous space word representations. In Proceedings of the 2013 conference of the north american chapter of the association for computational linguistics: Human language technologies. 746\u2013751."},{"key":"e_1_3_3_3_52_1","first-page":"2924","article-title":"On the Number of Linear Regions of Deep Neural Networks","volume":"27","author":"Montufar F","year":"2014","unstructured":"Guido\u00a0F Montufar, Razvan Pascanu, Kyunghyun Cho, and Yoshua Bengio. 2014. On the Number of Linear Regions of Deep Neural Networks. Advances in Neural Information Processing Systems 27 (2014), 2924\u20132932.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_3_3_53_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.2120481119"},{"key":"e_1_3_3_3_54_1","volume-title":"Neuro-Symbolic Reasoning. In Advances in Neural Information Processing Systems, M.\u00a0Ranzato, A.\u00a0Beygelzimer, Y.\u00a0Dauphin, P.S. Liang, and J.\u00a0Wortman Vaughan (Eds.). Vol.\u00a034. Curran Associates","author":"Nye Maxwell","year":"2021","unstructured":"Maxwell Nye, Michael Tessler, Josh Tenenbaum, and Brenden\u00a0M Lake. 2021. Improving Coherence and Consistency in Neural Sequence Models with Dual-System, Neuro-Symbolic Reasoning. In Advances in Neural Information Processing Systems, M.\u00a0Ranzato, A.\u00a0Beygelzimer, Y.\u00a0Dauphin, P.S. Liang, and J.\u00a0Wortman Vaughan (Eds.). Vol.\u00a034. Curran Associates, Inc., 25192\u201325204. https:\/\/proceedings.neurips.cc\/paper\/2021\/file\/d3e2e8f631bd9336ed25b8162aef8782-Paper.pdf"},{"key":"e_1_3_3_3_55_1","unstructured":"Catherine Olsson Nelson Elhage Neel Nanda Nicholas Joseph Nova DasSarma Tom Henighan Ben Mann Amanda Askell Yuntao Bai Anna Chen Tom Conerly Dawn Drain Deep Ganguli Zac Hatfield-Dodds Danny Hernandez Scott Johnston Andy Jones Jackson Kernion Liane Lovitt Kamal Ndousse Dario Amodei Tom Brown Jack Clark Jared Kaplan Sam McCandlish and Chris Olah. 2022. In-context Learning and Induction Heads. Transformer Circuits Thread(2022). https:\/\/transformer-circuits.pub\/2022\/in-context-learning-and-induction-heads\/index.html."},{"key":"e_1_3_3_3_56_1","unstructured":"OpenAI. 2022. ChatGPT: Optimizing Language Models for Dialogue. https:\/\/openai.com\/blog\/chatgpt\/"},{"key":"e_1_3_3_3_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3330340"},{"key":"e_1_3_3_3_58_1","doi-asserted-by":"publisher","unstructured":"Long Ouyang Jeff Wu Xu Jiang Diogo Almeida Carroll\u00a0L. Wainwright Pamela Mishkin Chong Zhang Sandhini Agarwal Katarina Slama Alex Ray John Schulman Jacob Hilton Fraser Kelton Luke Miller Maddie Simens Amanda Askell Peter Welinder Paul Christiano Jan Leike and Ryan Lowe. 2022. Training language models to follow instructions with human feedback. https:\/\/doi.org\/10.48550\/ARXIV.2203.02155","DOI":"10.48550\/ARXIV.2203.02155"},{"key":"e_1_3_3_3_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/3526113.3545616"},{"key":"e_1_3_3_3_60_1","first-page":"1","article-title":"Attention is Turing-Complete.J","volume":"22","author":"P\u00e9rez Jorge","year":"2021","unstructured":"Jorge P\u00e9rez, Pablo Barcel\u00f3, and Javier Marinkovic. 2021. Attention is Turing-Complete.J. Mach. Learn. Res. 22, 75 (2021), 1\u201335.","journal-title":"Mach. Learn. Res."},{"key":"e_1_3_3_3_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3174035"},{"key":"e_1_3_3_3_62_1","volume-title":"Exponential expressivity in deep neural networks through transient chaos. Advances in neural information processing systems 29","author":"Poole Ben","year":"2016","unstructured":"Ben Poole, Subhaneil Lahiri, Maithra Raghu, Jascha Sohl-Dickstein, and Surya Ganguli. 2016. Exponential expressivity in deep neural networks through transient chaos. Advances in neural information processing systems 29 (2016), 3360\u20133368."},{"key":"e_1_3_3_3_63_1","volume-title":"Grokking: Generalization beyond overfitting on small algorithmic datasets. arXiv preprint arXiv:2201.02177(2022).","author":"Power Alethea","year":"2022","unstructured":"Alethea Power, Yuri Burda, Harri Edwards, Igor Babuschkin, and Vedant Misra. 2022. Grokking: Generalization beyond overfitting on small algorithmic datasets. arXiv preprint arXiv:2201.02177(2022)."},{"key":"e_1_3_3_3_64_1","unstructured":"Alec Radford Rafal Jozefowicz and Ilya Sutskever. 2017. Learning to generate reviews and discovering sentiment. arXiv preprint arXiv:1704.01444(2017)."},{"key":"e_1_3_3_3_65_1","volume-title":"Language models are unsupervised multitask learners. OpenAI blog 1, 8","author":"Radford Alec","year":"2019","unstructured":"Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, 2019. Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019), 9."},{"key":"e_1_3_3_3_66_1","unstructured":"Jack\u00a0W. Rae Sebastian Borgeaud Trevor Cai Katie Millican Jordan Hoffmann Francis Song John Aslanides Sarah Henderson Roman Ring Susannah Young Eliza Rutherford Tom Hennigan Jacob Menick Albin Cassirer Richard Powell George van\u00a0den Driessche Lisa\u00a0Anne Hendricks Maribeth Rauh Po-Sen Huang Amelia Glaese Johannes Welbl Sumanth Dathathri Saffron Huang Jonathan Uesato John Mellor Irina Higgins Antonia Creswell Nat McAleese Amy Wu Erich Elsen Siddhant Jayakumar Elena Buchatskaya David Budden Esme Sutherland Karen Simonyan Michela Paganini Laurent Sifre Lena Martens Xiang\u00a0Lorraine Li Adhiguna Kuncoro Aida Nematzadeh Elena Gribovskaya Domenic Donato Angeliki Lazaridou Arthur Mensch Jean-Baptiste Lespiau Maria Tsimpoukelli Nikolai Grigorev Doug Fritz Thibault Sottiaux Mantas Pajarskas Toby Pohlen Zhitao Gong Daniel Toyama Cyprien de Masson\u00a0d\u2019Autume Yujia Li Tayfun Terzi Vladimir Mikulik Igor Babuschkin Aidan Clark Diego de Las\u00a0Casas Aurelia Guy Chris Jones James Bradbury Matthew Johnson Blake Hechtman Laura Weidinger Iason Gabriel William Isaac Ed Lockhart Simon Osindero Laura Rimell Chris Dyer Oriol Vinyals Kareem Ayoub Jeff Stanway Lorrayne Bennett Demis Hassabis Koray Kavukcuoglu and Geoffrey Irving. 2021. Scaling Language Models: Methods Analysis & Insights from Training Gopher. arxiv:2112.11446\u00a0[cs.CL]"},{"key":"e_1_3_3_3_67_1","article-title":"Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer","volume":"21","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter\u00a0J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 21(2020), 140:1\u2013140:67. http:\/\/jmlr.org\/papers\/v21\/20-074.html","journal-title":"J. Mach. Learn. Res."},{"key":"e_1_3_3_3_68_1","volume-title":"Proceedings of the 38th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a0139)","author":"Ramesh Aditya","year":"2021","unstructured":"Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. 2021. Zero-Shot Text-to-Image Generation. In Proceedings of the 38th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a0139), Marina Meila and Tong Zhang (Eds.). PMLR, 8821\u20138831. https:\/\/proceedings.mlr.press\/v139\/ramesh21a.html"},{"key":"e_1_3_3_3_69_1","doi-asserted-by":"publisher","DOI":"10.1145\/3410404.3414235"},{"key":"e_1_3_3_3_70_1","doi-asserted-by":"publisher","DOI":"10.1109\/5.880083"},{"key":"e_1_3_3_3_71_1","unstructured":"Victor Sanh Albert Webson Colin Raffel Stephen Bach Lintang Sutawika Zaid Alyafeai Antoine Chaffin Arnaud Stiegler Teven Scao Arun Raja 2022. Multitask Prompted Training Enables Zero-Shot Task Generalization. (2022)."},{"key":"e_1_3_3_3_72_1","doi-asserted-by":"crossref","unstructured":"Jeff Sauro and James\u00a0R Lewis. 2016. Quantifying the user experience: Practical statistics for user research. Morgan Kaufmann.","DOI":"10.1016\/B978-0-12-802308-2.00002-3"},{"key":"e_1_3_3_3_73_1","doi-asserted-by":"publisher","unstructured":"Stanislau Semeniuta Aliaksei Severyn and Sylvain Gelly. 2018. On Accurate Evaluation of GANs for Language Generation. https:\/\/doi.org\/10.48550\/ARXIV.1806.04936","DOI":"10.48550\/ARXIV.1806.04936"},{"key":"e_1_3_3_3_74_1","unstructured":"Tom Simonite. 2021. It Began as an AI-Fueled Dungeon Game. It Got Much Darker.https:\/\/www.wired.com\/story\/ai-fueled-dungeon-game-got-much-darker\/"},{"key":"e_1_3_3_3_75_1","unstructured":"Shaden Smith Mostofa Patwary Brandon Norick Patrick LeGresley Samyam Rajbhandari Jared Casper Zhun Liu Shrimai Prabhumoye George Zerveas Vijay Korthikanti Elton Zhang Rewon Child Reza\u00a0Yazdani Aminabadi Julie Bernauer Xia Song Mohammad Shoeybi Yuxiong He Michael Houston Saurabh Tiwary and Bryan Catanzaro. 2022. Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B A Large-Scale Generative Language Model. arxiv:2201.11990\u00a0[cs.CL]"},{"key":"e_1_3_3_3_76_1","doi-asserted-by":"publisher","DOI":"10.1145\/3490100.3516464"},{"key":"e_1_3_3_3_77_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan\u00a0N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information processing systems. 5998\u20136008."},{"key":"e_1_3_3_3_78_1","unstructured":"Jason Wei Yi Tay Rishi Bommasani Colin Raffel Barret Zoph Sebastian Borgeaud Dani Yogatama Maarten Bosma Denny Zhou Donald Metzler 2022. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682(2022)."},{"key":"e_1_3_3_3_79_1","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.2201.11903"},{"key":"e_1_3_3_3_80_1","doi-asserted-by":"publisher","DOI":"10.1145\/1462198.1462203"},{"key":"e_1_3_3_3_81_1","volume-title":"Crossfit: A few-shot learning challenge for cross-task generalization in nlp. arXiv preprint arXiv:2104.08835(2021).","author":"Ye Qinyuan","year":"2021","unstructured":"Qinyuan Ye, Bill\u00a0Yuchen Lin, and Xiang Ren. 2021. Crossfit: A few-shot learning challenge for cross-task generalization in nlp. arXiv preprint arXiv:2104.08835(2021)."},{"key":"e_1_3_3_3_82_1","doi-asserted-by":"publisher","DOI":"10.4324\/9781003025245-12"},{"key":"e_1_3_3_3_83_1","volume-title":"Proceedings of the 38th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a0139)","author":"Zhao Zihao","year":"2021","unstructured":"Zihao Zhao, Eric Wallace, Shi Feng, Dan Klein, and Sameer Singh. 2021. Calibrate Before Use: Improving Few-shot Performance of Language Models. In Proceedings of the 38th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a0139), Marina Meila and Tong Zhang (Eds.). PMLR, 12697\u201312706. https:\/\/proceedings.mlr.press\/v139\/zhao21c.html"}],"event":{"name":"CHI '23: CHI Conference on Human Factors in Computing Systems","location":"Hamburg Germany","acronym":"CHI '23","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3544548.3580688","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3544548.3580688","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:23Z","timestamp":1750178243000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3544548.3580688"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,19]]},"references-count":83,"alternative-id":["10.1145\/3544548.3580688","10.1145\/3544548"],"URL":"https:\/\/doi.org\/10.1145\/3544548.3580688","relation":{},"subject":[],"published":{"date-parts":[[2023,4,19]]},"assertion":[{"value":"2023-04-19","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}