{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,30]],"date-time":"2026-04-30T17:34:39Z","timestamp":1777570479946,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":34,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,4,12]],"date-time":"2023-04-12T00:00:00Z","timestamp":1681257600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,4,12]]},"DOI":"10.1145\/3582437.3587211","type":"proceedings-article","created":{"date-parts":[[2023,4,11]],"date-time":"2023-04-11T04:13:33Z","timestamp":1681186413000},"page":"1-8","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":75,"title":["Level Generation Through Large Language Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-8194-8854","authenticated-orcid":false,"given":"Graham","family":"Todd","sequence":"first","affiliation":[{"name":"New York University Tandon, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3783-9486","authenticated-orcid":false,"given":"Sam","family":"Earle","sequence":"additional","affiliation":[{"name":"New York University Tandon, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2458-9599","authenticated-orcid":false,"given":"Muhammad Umair","family":"Nasir","sequence":"additional","affiliation":[{"name":"University of the Witwatersrand, South Africa"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3366-8165","authenticated-orcid":false,"given":"Michael Cerny","family":"Green","sequence":"additional","affiliation":[{"name":"New York University Tandon, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3128-4598","authenticated-orcid":false,"given":"Julian","family":"Togelius","sequence":"additional","affiliation":[{"name":"New York University Tandon, United States"}]}],"member":"320","published-online":{"date-parts":[[2023,4,12]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"MusicLM: Generating Music From Text. arXiv preprint arXiv:2301.11325","author":"Agostinelli Andrea","year":"2023","unstructured":"Andrea Agostinelli, Timo\u00a0I Denk, Zal\u00e1n Borsos, Jesse Engel, Mauro Verzetti, Antoine Caillon, Qingqing Huang, Aren Jansen, Adam Roberts, Marco Tagliasacchi, 2023. MusicLM: Generating Music From Text. arXiv preprint arXiv:2301.11325 (2023)."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","unstructured":"Tom\u00a0B. Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell Sandhini Agarwal Ariel Herbert-Voss Gretchen Krueger Tom Henighan Rewon Child Aditya Ramesh Daniel\u00a0M. Ziegler Jeffrey Wu Clemens Winter Christopher Hesse Mark Chen Eric Sigler Mateusz Litwin Scott Gray Benjamin Chess Jack Clark Christopher Berner Sam McCandlish Alec Radford Ilya Sutskever and Dario Amodei. 2020. Language Models are Few-Shot Learners. https:\/\/doi.org\/10.48550\/ARXIV.2005.14165","DOI":"10.48550\/ARXIV.2005.14165"},{"key":"e_1_3_2_1_3_1","volume-title":"Decision transformer: Reinforcement learning via sequence modeling. Advances in neural information processing systems 34","author":"Chen Lili","year":"2021","unstructured":"Lili Chen, Kevin Lu, Aravind Rajeswaran, Kimin Lee, Aditya Grover, Misha Laskin, Pieter Abbeel, Aravind Srinivas, and Igor Mordatch. 2021. Decision transformer: Reinforcement learning via sequence modeling. Advances in neural information processing systems 34 (2021), 15084\u201315097."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/2676467.2676506"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/CoG52621.2021.9619159"},{"key":"e_1_3_2_1_6_1","volume-title":"Music transformer. arXiv preprint arXiv:1809.04281","author":"Huang Zhi\u00a0Anna","year":"2018","unstructured":"Cheng-Zhi\u00a0Anna Huang, Ashish Vaswani, Jakob Uszkoreit, Noam Shazeer, Ian Simon, Curtis Hawthorne, Andrew\u00a0M Dai, Matthew\u00a0D Hoffman, Monica Dinculescu, and Douglas Eck. 2018. Music transformer. arXiv preprint arXiv:1809.04281 (2018)."},{"key":"e_1_3_2_1_7_1","volume-title":"Scaling laws for neural language models. arXiv preprint arXiv:2001.08361","author":"Kaplan Jared","year":"2020","unstructured":"Jared Kaplan, Sam McCandlish, Tom Henighan, Tom\u00a0B Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. 2020. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361 (2020)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.5555\/3505288.3505297"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3337722.3341845"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1609\/aiide.v16i1.7416"},{"key":"e_1_3_2_1_11_1","volume-title":"Learning to decode for future success. arXiv preprint arXiv:1701.06549","author":"Li Jiwei","year":"2017","unstructured":"Jiwei Li, Will Monroe, and Dan Jurafsky. 2017. Learning to decode for future success. arXiv preprint arXiv:1701.06549 (2017)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00521-020-05383-8"},{"key":"e_1_3_2_1_13_1","volume-title":"Codexglue: A machine learning benchmark dataset for code understanding and generation. arXiv preprint arXiv:2102.04664","author":"Lu Shuai","year":"2021","unstructured":"Shuai Lu, Daya Guo, Shuo Ren, Junjie Huang, Alexey Svyatkovskiy, Ambrosio Blanco, Colin Clement, Dawn Drain, Daxin Jiang, Duyu Tang, 2021. Codexglue: A machine learning benchmark dataset for code understanding and generation. arXiv preprint arXiv:2102.04664 (2021)."},{"key":"e_1_3_2_1_14_1","unstructured":"Justin Mott Saujas Nandi and Luke Zeller. 2019. Controllable and coherent level generation: A two-pronged approach. In Experimental AI in games workshop."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1007\/3-540-61532-6_50"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/CIG.2019.8848085"},{"key":"e_1_3_2_1_17_1","volume-title":"Language models are unsupervised multitask learners. OpenAI blog 1, 8","author":"Radford Alec","year":"2019","unstructured":"Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, 2019. Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019), 9."},{"key":"e_1_3_2_1_18_1","volume-title":"Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125","author":"Ramesh Aditya","year":"2022","unstructured":"Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. 2022. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125 (2022)."},{"key":"e_1_3_2_1_19_1","volume-title":"A generalist agent. arXiv preprint arXiv:2205.06175","author":"Reed Scott","year":"2022","unstructured":"Scott Reed, Konrad Zolna, Emilio Parisotto, Sergio\u00a0Gomez Colmenarejo, Alexander Novikov, Gabriel Barth-Maron, Mai Gimenez, Yury Sulsky, Jackie Kay, Jost\u00a0Tobias Springenberg, 2022. A generalist agent. arXiv preprint arXiv:2205.06175 (2022)."},{"key":"e_1_3_2_1_20_1","volume-title":"AIIDE Workshops.","author":"Sarkar Anurag","year":"2018","unstructured":"Anurag Sarkar and Seth Cooper. 2018. Blending Levels from Different Games using LSTMs.. In AIIDE Workshops."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/CoG52621.2021.9619051"},{"key":"e_1_3_2_1_22_1","volume-title":"Conditional level generation and game blending. arXiv preprint arXiv:2010.07735","author":"Sarkar Anurag","year":"2020","unstructured":"Anurag Sarkar, Zhihan Yang, and Seth Cooper. 2020. Conditional level generation and game blending. arXiv preprint arXiv:2010.07735 (2020)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377930.3389821"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3402942.3402948"},{"key":"e_1_3_2_1_25_1","volume-title":"MarioGPT: Open-Ended Text2Level Generation through Large Language Models. arXiv preprint arXiv:2302.05981","author":"Sudhakaran Shyam","year":"2023","unstructured":"Shyam Sudhakaran, Miguel Gonz\u00e1lez-Duque, Claire Glanois, Matthias Freiberger, Elias Najarro, and Sebastian Risi. 2023. MarioGPT: Open-Ended Text2Level Generation through Large Language Models. arXiv preprint arXiv:2302.05981 (2023)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.14569\/IJACSA.2017.080364"},{"key":"e_1_3_2_1_27_1","unstructured":"Adam Summerville Matthew Guzdial Michael Mateas and Mark\u00a0O Riedl. 2016. Learning player tailored content from observation: Platformer level generation from video traces using lstms. In Twelfth artificial intelligence and interactive digital entertainment conference."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","unstructured":"Adam Summerville and Michael Mateas. 2016. Super Mario as a String: Platformer Level Generation Via LSTMs. https:\/\/doi.org\/10.48550\/ARXIV.1603.00930","DOI":"10.48550\/ARXIV.1603.00930"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/TG.2018.2846639"},{"key":"e_1_3_2_1_30_1","volume-title":"Proceedings of the International North American Conference on Intelligent Games and Simulation. 5\u201312","author":"Taylor Joshua","year":"2011","unstructured":"Joshua Taylor and Ian Parberry. 2011. Procedural generation of sokoban levels. In Proceedings of the International North American Conference on Intelligent Games and Simulation. 5\u201312."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/CoG47356.2020.9231576"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","unstructured":"Th\u00e9ophane Weber S\u00e9bastien Racani\u00e8re David\u00a0P. Reichert Lars Buesing Arthur Guez Danilo\u00a0Jimenez Rezende Adria\u00a0Puigdom\u00e8nech Badia Oriol Vinyals Nicolas Heess Yujia Li Razvan Pascanu Peter Battaglia Demis Hassabis David Silver and Daan Wierstra. 2017. Imagination-Augmented Agents for Deep Reinforcement Learning. https:\/\/doi.org\/10.48550\/ARXIV.1707.06203","DOI":"10.48550\/ARXIV.1707.06203"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/TG.2022.3175795"},{"key":"e_1_3_2_1_34_1","volume-title":"International Conference on Machine Learning. PMLR, 12697\u201312706","author":"Zhao Zihao","year":"2021","unstructured":"Zihao Zhao, Eric Wallace, Shi Feng, Dan Klein, and Sameer Singh. 2021. Calibrate before use: Improving few-shot performance of language models. In International Conference on Machine Learning. PMLR, 12697\u201312706."}],"event":{"name":"FDG 2023: Foundations of Digital Games 2023","location":"Lisbon Portugal","acronym":"FDG 2023"},"container-title":["Proceedings of the 18th International Conference on the Foundations of Digital Games"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3582437.3587211","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3582437.3587211","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T18:08:50Z","timestamp":1750183730000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3582437.3587211"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,12]]},"references-count":34,"alternative-id":["10.1145\/3582437.3587211","10.1145\/3582437"],"URL":"https:\/\/doi.org\/10.1145\/3582437.3587211","relation":{},"subject":[],"published":{"date-parts":[[2023,4,12]]},"assertion":[{"value":"2023-04-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}