{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T07:33:07Z","timestamp":1776238387467,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":85,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,3,18]],"date-time":"2024-03-18T00:00:00Z","timestamp":1710720000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,3,18]]},"DOI":"10.1145\/3640543.3645143","type":"proceedings-article","created":{"date-parts":[[2024,4,5]],"date-time":"2024-04-05T18:23:12Z","timestamp":1712341392000},"page":"699-714","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":60,"title":["LAVE: LLM-Powered Agent Assistance and Language Augmentation for Video Editing"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9016-038X","authenticated-orcid":false,"given":"Bryan","family":"Wang","sequence":"first","affiliation":[{"name":"Department of Computer Science, University of Toronto, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0602-149X","authenticated-orcid":false,"given":"Yuliang","family":"Li","sequence":"additional","affiliation":[{"name":"Reality Lab Research, Meta, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7788-9982","authenticated-orcid":false,"given":"Zhaoyang","family":"Lv","sequence":"additional","affiliation":[{"name":"Meta Reality Labs Research, Meta, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9425-0881","authenticated-orcid":false,"given":"Haijun","family":"Xia","sequence":"additional","affiliation":[{"name":"Department of Cognitive Science and Design Lab, University of California, San Diego, United States"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-3890-9847","authenticated-orcid":false,"given":"Yan","family":"Xu","sequence":"additional","affiliation":[{"name":"Reality Labs Research, Meta, United States"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-8729-1791","authenticated-orcid":false,"given":"Raj","family":"Sodhi","sequence":"additional","affiliation":[{"name":"Facebook Reality Labs, Facebook, United States"}]}],"member":"320","published-online":{"date-parts":[[2024,4,5]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"2023. Adobe Premiere Pro. https:\/\/www.adobe.com\/products\/premiere.html"},{"key":"e_1_3_2_1_2_1","unstructured":"2023. ChromaDB. https:\/\/www.trychroma.com\/"},{"key":"e_1_3_2_1_3_1","unstructured":"2023. Final Cut Pro. https:\/\/www.apple.com\/final-cut-pro\/"},{"key":"e_1_3_2_1_4_1","unstructured":"2023. Function calling and other API updates. https:\/\/openai.com\/blog\/function-calling-and-other-api-updates"},{"key":"e_1_3_2_1_5_1","unstructured":"2023. Gen-2 Runway. https:\/\/runwayml.com\/ai-magic-tools\/gen-2\/"},{"key":"e_1_3_2_1_6_1","unstructured":"2023. Langchain. https:\/\/www.langchain.com\/"},{"key":"e_1_3_2_1_7_1","volume-title":"Large language models are zero-shot clinical information extractors. arXiv preprint arXiv:2205.12689","author":"Agrawal Monica","year":"2022","unstructured":"Monica Agrawal, Stefan Hegselmann, Hunter Lang, Yoon Kim, and David Sontag. 2022. Large language models are zero-shot clinical information extractors. arXiv preprint arXiv:2205.12689 (2022)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300233"},{"key":"e_1_3_2_1_9_1","volume-title":"ChainForge: A Visual Toolkit for Prompt Engineering and LLM Hypothesis Testing. arXiv preprint arXiv:2309.09128","author":"Arawjo Ian","year":"2023","unstructured":"Ian Arawjo, Chelse Swoopes, Priyan Vaithilingam, Martin Wattenberg, and Elena Glassman. 2023. ChainForge: A Visual Toolkit for Prompt Engineering and LLM Hypothesis Testing. arXiv preprint arXiv:2309.09128 (2023)."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1080\/19393555.2022.2060879"},{"key":"e_1_3_2_1_11_1","volume-title":"Promptify: Text-to-Image Generation through Interactive Prompt Exploration with Large Language Models. arxiv:2304.09337\u00a0[cs.HC]","author":"Brade Stephen","year":"2023","unstructured":"Stephen Brade, Bryan Wang, Mauricio Sousa, Sageev Oore, and Tovi Grossman. 2023. Promptify: Text-to-Image Generation through Interactive Prompt Exploration with Large Language Models. arxiv:2304.09337\u00a0[cs.HC]"},{"key":"e_1_3_2_1_12_1","volume-title":"ChemCrow: Augmenting large-language models with chemistry tools. arXiv preprint arXiv:2304.05376","author":"Bran M","year":"2023","unstructured":"Andres\u00a0M Bran, Sam Cox, Andrew\u00a0D White, and Philippe Schwaller. 2023. ChemCrow: Augmenting large-language models with chemistry tools. arXiv preprint arXiv:2304.05376 (2023)."},{"key":"e_1_3_2_1_13_1","unstructured":"Tom\u00a0B. Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell Sandhini Agarwal Ariel Herbert-Voss Gretchen Krueger Tom Henighan Rewon Child Aditya Ramesh Daniel\u00a0M. Ziegler Jeffrey Wu Clemens Winter Christopher Hesse Mark Chen Eric Sigler Mateusz Litwin Scott Gray Benjamin Chess Jack Clark Christopher Berner Sam McCandlish Alec Radford Ilya Sutskever and Dario Amodei. 2020. Language Models are Few-Shot Learners. arxiv:2005.14165\u00a0[cs.CL]"},{"key":"e_1_3_2_1_14_1","volume-title":"Nine potential pitfalls when designing human-ai co-creative systems. arXiv preprint arXiv:2104.00358","author":"Buschek Daniel","year":"2021","unstructured":"Daniel Buschek, Lukas Mecke, Florian Lehmann, and Hai Dang. 2021. Nine potential pitfalls when designing human-ai co-creative systems. arXiv preprint arXiv:2104.00358 (2021)."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"crossref","unstructured":"Tuhin Chakrabarty Vishakh Padmakumar Faeze Brahman and Smaranda Muresan. 2023. Creativity Support in the Age of Large Language Models: An Empirical Study Involving Emerging Writers. arxiv:2309.12570\u00a0[cs.HC]","DOI":"10.1145\/3635636.3656201"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445131"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300931"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/2617588"},{"key":"e_1_3_2_1_19_1","volume-title":"Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality. https:\/\/lmsys.org\/blog\/2023-03-30-vicuna\/","author":"Chiang Wei-Lin","year":"2023","unstructured":"Wei-Lin Chiang, Zhuohan Li, Zi Lin, Ying Sheng, Zhanghao Wu, Hao Zhang, Lianmin Zheng, Siyuan Zhuang, Yonghao Zhuang, Joseph\u00a0E. Gonzalez, Ion Stoica, and Eric\u00a0P. Xing. 2023. Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality. https:\/\/lmsys.org\/blog\/2023-03-30-vicuna\/"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3501819"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-020-0161-x"},{"key":"e_1_3_2_1_22_1","volume-title":"WinoQueer: A Community-in-the-Loop Benchmark for Anti-LGBTQ+ Bias in Large Language Models. arXiv preprint arXiv:2306.15087","author":"Felkner K","year":"2023","unstructured":"Virginia\u00a0K Felkner, Ho-Chun\u00a0Herbert Chang, Eugene Jang, and Jonathan May. 2023. WinoQueer: A Community-in-the-Loop Benchmark for Anti-LGBTQ+ Bias in Large Language Models. arXiv preprint arXiv:2306.15087 (2023)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306346.3323028"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.5465\/annals.2018.0057"},{"key":"e_1_3_2_1_25_1","volume-title":"Imagen Video: High Definition Video Generation with Diffusion Models. arxiv:2210.02303\u00a0[cs.CV]","author":"Ho Jonathan","year":"2022","unstructured":"Jonathan Ho, William Chan, Chitwan Saharia, Jay Whang, Ruiqi Gao, Alexey Gritsenko, Diederik\u00a0P. Kingma, Ben Poole, Mohammad Norouzi, David\u00a0J. Fleet, and Tim Salimans. 2022. Imagen Video: High Definition Video Generation with Diffusion Models. arxiv:2210.02303\u00a0[cs.CV]"},{"key":"e_1_3_2_1_26_1","volume-title":"Monica Dinculescu, and Carrie\u00a0J Cai.","author":"Huang Zhi\u00a0Anna","year":"2020","unstructured":"Cheng-Zhi\u00a0Anna Huang, Hendrik\u00a0Vincent Koops, Ed Newton-Rex, Monica Dinculescu, and Carrie\u00a0J Cai. 2020. AI song contest: Human-AI co-creation in songwriting. arXiv preprint arXiv:2010.05388 (2020)."},{"key":"e_1_3_2_1_27_1","unstructured":"Wenlong Huang Pieter Abbeel Deepak Pathak and Igor Mordatch. 2022. Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents. arxiv:2201.07207\u00a0[cs.LG]"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300311"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581494"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1093\/jcmc\/zmac014"},{"key":"e_1_3_2_1_31_1","unstructured":"Ehud Karpas Omri Abend Yonatan Belinkov Barak Lenz Opher Lieber Nir Ratner Yoav Shoham Hofit Bata Yoav Levine Kevin Leyton-Brown Dor Muhlgay Noam Rozen Erez Schwartz Gal Shachaf Shai Shalev-Shwartz Amnon Shashua and Moshe Tenenholtz. 2022. MRKL Systems: A modular neuro-symbolic architecture that combines large language models external knowledge sources and discrete reasoning. arxiv:2205.00445\u00a0[cs.CL]"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3415234"},{"key":"e_1_3_2_1_33_1","volume-title":"EvalLM: Interactive Evaluation of Large Language Model Prompts on User-Defined Criteria. arXiv preprint arXiv:2309.13633","author":"Kim Tae\u00a0Soo","year":"2023","unstructured":"Tae\u00a0Soo Kim, Yoonjoo Lee, Jamin Shin, Young-Ho Kim, and Juho Kim. 2023. EvalLM: Interactive Evaluation of Large Language Model Prompts on User-Defined Criteria. arXiv preprint arXiv:2309.13633 (2023)."},{"key":"e_1_3_2_1_34_1","unstructured":"Takeshi Kojima Shixiang\u00a0Shane Gu Machel Reid Yutaka Matsuo and Yusuke Iwasawa. 2023. Large Language Models are Zero-Shot Reasoners. arxiv:2205.11916\u00a0[cs.CL]"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/2470654.2481301"},{"key":"e_1_3_2_1_36_1","unstructured":"Belinda\u00a0Z. Li Maxwell Nye and Jacob Andreas. 2021. Implicit Representations of Meaning in Neural Language Models. arxiv:2106.00737\u00a0[cs.CL]"},{"key":"e_1_3_2_1_37_1","unstructured":"Junnan Li Dongxu Li Silvio Savarese and Steven Hoi. 2023. BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models. arxiv:2301.12597\u00a0[cs.CV]"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581006"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1093\/jcmc\/zmab013"},{"key":"e_1_3_2_1_40_1","unstructured":"Haotian Liu Chunyuan Li Qingyang Wu and Yong\u00a0Jae Lee. 2023. Visual Instruction Tuning. (2023)."},{"key":"e_1_3_2_1_41_1","volume-title":"Generative Disco: Text-to-Video Generation for Music Visualization. arXiv preprint arXiv:2304.08551","author":"Liu Vivian","year":"2023","unstructured":"Vivian Liu, Tao Long, Nathan Raw, and Lydia Chilton. 2023. Generative Disco: Text-to-Video Generation for Music Visualization. arXiv preprint arXiv:2304.08551 (2023)."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3526113.3545621"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3563657.3596098"},{"key":"e_1_3_2_1_44_1","volume-title":"Cutting down on prompts and parameters: Simple few-shot learning with language models. arXiv preprint arXiv:2106.13353","author":"L","year":"2021","unstructured":"Robert\u00a0L Logan\u00a0IV, Ivana Bala\u017eevi\u0107, Eric Wallace, Fabio Petroni, Sameer Singh, and Sebastian Riedel. 2021. Cutting down on prompts and parameters: Simple few-shot learning with language models. arXiv preprint arXiv:2106.13353 (2021)."},{"key":"e_1_3_2_1_45_1","unstructured":"R\u00f3is\u00edn Loughran. 2022. Bias and Creativity.. In ICCC. 354\u2013358."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376739"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3490099.3511159"},{"key":"e_1_3_2_1_48_1","volume-title":"Humans as Creativity Gatekeepers: Are We Biased Against AI Creativity?Journal of Business and Psychology","author":"Magni Federico","year":"2023","unstructured":"Federico Magni, Jiyoung Park, and Melody\u00a0Manchi Chao. 2023. Humans as Creativity Gatekeepers: Are We Biased Against AI Creativity?Journal of Business and Psychology (2023), 1\u201314."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-16667-0_3"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581225"},{"key":"e_1_3_2_1_51_1","volume-title":"David Bieber, David Dohan, Aitor Lewkowycz, Maarten Bosma, David Luan, Charles Sutton, and Augustus Odena.","author":"Nye Maxwell","year":"2021","unstructured":"Maxwell Nye, Anders\u00a0Johan Andreassen, Guy Gur-Ari, Henryk Michalewski, Jacob Austin, David Bieber, David Dohan, Aitor Lewkowycz, Maarten Bosma, David Luan, Charles Sutton, and Augustus Odena. 2021. Show Your Work: Scratchpads for Intermediate Computation with Language Models. arxiv:2112.00114\u00a0[cs.LG]"},{"key":"e_1_3_2_1_52_1","unstructured":"OpenAI. 2023. GPT-4 Technical Report. arxiv:2303.08774\u00a0[cs.CL]"},{"key":"e_1_3_2_1_53_1","volume-title":"Generative agents: Interactive simulacra of human behavior. arXiv preprint arXiv:2304.03442","author":"Park Joon\u00a0Sung","year":"2023","unstructured":"Joon\u00a0Sung Park, Joseph\u00a0C O\u2019Brien, Carrie\u00a0J Cai, Meredith\u00a0Ringel Morris, Percy Liang, and Michael\u00a0S Bernstein. 2023. Generative agents: Interactive simulacra of human behavior. arXiv preprint arXiv:2304.03442 (2023)."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/3311957.3359433"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/3379337.3415864"},{"key":"e_1_3_2_1_56_1","volume-title":"Identifying ethical issues in ai partners in human-ai co-creation. arXiv preprint arXiv:2204.07644","author":"Rezwana Jeba","year":"2022","unstructured":"Jeba Rezwana and Mary\u00a0Lou Maher. 2022. Identifying ethical issues in ai partners in human-ai co-creation. arXiv preprint arXiv:2204.07644 (2022)."},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.437"},{"key":"e_1_3_2_1_58_1","volume-title":"From Pixels to UI Actions: Learning to Follow Instructions via Graphical User Interfaces. arXiv preprint arXiv:2306.00245","author":"Shaw Peter","year":"2023","unstructured":"Peter Shaw, Mandar Joshi, James Cohan, Jonathan Berant, Panupong Pasupat, Hexiang Hu, Urvashi Khandelwal, Kenton Lee, and Kristina Toutanova. 2023. From Pixels to UI Actions: Learning to Follow Instructions via Graphical User Interfaces. arXiv preprint arXiv:2306.00245 (2023)."},{"key":"e_1_3_2_1_59_1","unstructured":"Yongliang Shen Kaitao Song Xu Tan Dongsheng Li Weiming Lu and Yueting Zhuang. 2023. HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in Hugging Face. arxiv:2303.17580\u00a0[cs.CL]"},{"key":"e_1_3_2_1_60_1","unstructured":"Yusuxke Shibata Takuya Kida Shuichi Fukamachi Masayuki Takeda Ayumi Shinohara Takeshi Shinohara and Setsuo Arikawa. 1999. Byte Pair encoding: A text compression scheme that accelerates pattern matching. (1999)."},{"key":"e_1_3_2_1_61_1","volume-title":"Reflexion: an autonomous agent with dynamic memory and self-reflection. arXiv preprint arXiv:2303.11366","author":"Shinn Noah","year":"2023","unstructured":"Noah Shinn, Beck Labash, and Ashwin Gopinath. 2023. Reflexion: an autonomous agent with dynamic memory and self-reflection. arXiv preprint arXiv:2303.11366 (2023)."},{"key":"e_1_3_2_1_62_1","unstructured":"Uriel Singer Adam Polyak Thomas Hayes Xi Yin Jie An Songyang Zhang Qiyuan Hu Harry Yang Oron Ashual Oran Gafni Devi Parikh Sonal Gupta and Yaniv Taigman. 2022. Make-A-Video: Text-to-Video Generation without Text-Video Data. arxiv:2209.14792\u00a0[cs.CV]"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00280"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1145\/3526113.3545617"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"crossref","unstructured":"Derek Tam Anisha Mascarenhas Shiyue Zhang Sarah Kwan Mohit Bansal and Colin Raffel. 2022. Evaluating the Factual Consistency of Large Language Models Through Summarization. arxiv:2211.08412\u00a0[cs.CL]","DOI":"10.18653\/v1\/2023.findings-acl.322"},{"key":"e_1_3_2_1_66_1","unstructured":"Hugo Touvron Thibaut Lavril Gautier Izacard Xavier Martinet Marie-Anne Lachaux Timoth\u00e9e Lacroix Baptiste Rozi\u00e8re Naman Goyal Eric Hambro Faisal Azhar Aurelien Rodriguez Armand Joulin Edouard Grave and Guillaume Lample. 2023. LLaMA: Open and Efficient Foundation Language Models. arxiv:2302.13971\u00a0[cs.CL]"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1145\/2984511.2984569"},{"key":"e_1_3_2_1_68_1","volume-title":"The think aloud method: a practical approach to modelling cognitive","author":"Van\u00a0Someren Maarten","year":"1994","unstructured":"Maarten Van\u00a0Someren, Yvonne\u00a0F Barnard, and J Sandberg. 1994. The think aloud method: a practical approach to modelling cognitive. London: AcademicPress 11 (1994), 29\u201341."},{"key":"e_1_3_2_1_69_1","volume-title":"Nationality Bias in Text Generation. arXiv preprint arXiv:2302.02463","author":"Venkit Pranav\u00a0Narayanan","year":"2023","unstructured":"Pranav\u00a0Narayanan Venkit, Sanjana Gautam, Ruchi Panchanadikar, Shomir Wilson, 2023. Nationality Bias in Text Generation. arXiv preprint arXiv:2302.02463 (2023)."},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1145\/3526113.3545680"},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3580895"},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359313"},{"key":"e_1_3_2_1_73_1","volume-title":"Roy Ka-Wei Lee, and Ee-Peng Lim","author":"Wang Lei","year":"2023","unstructured":"Lei Wang, Wanyu Xu, Yihuai Lan, Zhiqiang Hu, Yunshi Lan, Roy Ka-Wei Lee, and Ee-Peng Lim. 2023. Plan-and-Solve Prompting: Improving Zero-Shot Chain-of-Thought Reasoning by Large Language Models. arxiv:2305.04091\u00a0[cs.CL]"},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1145\/3355089.3356520"},{"key":"e_1_3_2_1_75_1","volume-title":"ReelFramer: Co-creating News Reels on Social Media with Generative AI. arXiv preprint arXiv:2304.09653","author":"Wang Sitong","year":"2023","unstructured":"Sitong Wang, Samia Menon, Tao Long, Keren Henderson, Dingzeyu Li, Kevin Crowston, Mark Hansen, Jeffrey\u00a0V Nickerson, and Lydia\u00a0B Chilton. 2023. ReelFramer: Co-creating News Reels on Social Media with Generative AI. arXiv preprint arXiv:2304.09653 (2023)."},{"key":"e_1_3_2_1_76_1","volume-title":"Chi, Quoc Le, and Denny Zhou","author":"Wei Jason","year":"2023","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed Chi, Quoc Le, and Denny Zhou. 2023. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. arxiv:2201.11903\u00a0[cs.CL]"},{"key":"e_1_3_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1145\/3379337.3415845"},{"key":"e_1_3_2_1_78_1","doi-asserted-by":"publisher","DOI":"10.1145\/3379337.3415882"},{"key":"e_1_3_2_1_79_1","volume-title":"The dawn of lmms: Preliminary explorations with gpt-4v (ision). arXiv preprint arXiv:2309.17421 9, 1","author":"Yang Zhengyuan","year":"2023","unstructured":"Zhengyuan Yang, Linjie Li, Kevin Lin, Jianfeng Wang, Chung-Ching Lin, Zicheng Liu, and Lijuan Wang. 2023. The dawn of lmms: Preliminary explorations with gpt-4v (ision). arXiv preprint arXiv:2309.17421 9, 1 (2023)."},{"key":"e_1_3_2_1_80_1","unstructured":"Shunyu Yao Jeffrey Zhao Dian Yu Nan Du Izhak Shafran Karthik Narasimhan and Yuan Cao. 2023. ReAct: Synergizing Reasoning and Acting in Language Models. arxiv:2210.03629\u00a0[cs.CL]"},{"key":"e_1_3_2_1_81_1","volume-title":"Large Language Model as Attributed Training Data Generator: A Tale of Diversity and Bias. arXiv preprint arXiv:2306.15895","author":"Yu Yue","year":"2023","unstructured":"Yue Yu, Yuchen Zhuang, Jieyu Zhang, Yu Meng, Alexander Ratner, Ranjay Krishna, Jiaming Shen, and Chao Zhang. 2023. Large Language Model as Attributed Training Data Generator: A Tale of Diversity and Bias. arXiv preprint arXiv:2306.15895 (2023)."},{"key":"e_1_3_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1145\/3490099.3511105"},{"key":"e_1_3_2_1_83_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581388"},{"key":"e_1_3_2_1_84_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3517479"},{"key":"e_1_3_2_1_85_1","doi-asserted-by":"publisher","DOI":"10.1109\/CIG.2018.8490433"}],"event":{"name":"IUI '24: 29th International Conference on Intelligent User Interfaces","location":"Greenville SC USA","acronym":"IUI '24","sponsor":["SIGAI ACM Special Interest Group on Artificial Intelligence","SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Proceedings of the 29th International Conference on Intelligent User Interfaces"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3640543.3645143","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3640543.3645143","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:56:29Z","timestamp":1764550589000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3640543.3645143"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3,18]]},"references-count":85,"alternative-id":["10.1145\/3640543.3645143","10.1145\/3640543"],"URL":"https:\/\/doi.org\/10.1145\/3640543.3645143","relation":{},"subject":[],"published":{"date-parts":[[2024,3,18]]},"assertion":[{"value":"2024-04-05","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}