{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,16]],"date-time":"2025-07-16T13:16:47Z","timestamp":1752671807844,"version":"3.41.0"},"reference-count":93,"publisher":"Association for Computing Machinery (ACM)","issue":"3","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Comput.-Hum. Interact."],"published-print":{"date-parts":[[2025,6,30]]},"abstract":"<jats:p>UI task automation enables efficient task execution by simulating human interactions with GUIs, without modifying the existing application code. However, its broader adoption is constrained by the need for expertise in both scripting languages and workflow design. To address this challenge, we present Prompt2Task, a system designed to comprehend various task-related textual prompts (e.g., goals, procedures), thereby generating and performing the corresponding automation tasks. Prompt2Task incorporates a suite of intelligent agents that mimic human cognitive functions, specializing in interpreting user intent, managing external information for task generation, and executing operations on smartphones. The agents can learn from user feedback and continuously improve their performance based on the accumulated knowledge. Experimental results indicated a performance jump from a 22.28% success rate in the baseline to 95.24% with Prompt2Task, requiring an average of 0.69 user interventions for each new task. Prompt2Task presents promising applications in fields such as tutorial creation, smart assistance, and customer service.<\/jats:p>","DOI":"10.1145\/3716132","type":"journal-article","created":{"date-parts":[[2025,2,5]],"date-time":"2025-02-05T16:38:57Z","timestamp":1738773537000},"page":"1-45","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Prompt2Task: Automating UI Tasks on Smartphones from Textual Prompts"],"prefix":"10.1145","volume":"32","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-8639-5929","authenticated-orcid":false,"given":"Tian","family":"Huang","sequence":"first","affiliation":[{"name":"Department of Computer Science and Technology, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2591-7993","authenticated-orcid":false,"given":"Chun","family":"Yu","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1351-9034","authenticated-orcid":false,"given":"Weinan","family":"Shi","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-6971-0064","authenticated-orcid":false,"given":"Zijian","family":"Peng","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8260-3830","authenticated-orcid":false,"given":"David","family":"Yang","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-0957-4864","authenticated-orcid":false,"given":"Weiqi","family":"Sun","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Technology, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2273-6927","authenticated-orcid":false,"given":"Yuanchun","family":"Shi","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Tsinghua University, Beijing, China and Department of Computer Science, Qinghai University, Xining, China"}]}],"member":"320","published-online":{"date-parts":[[2025,6,14]]},"reference":[{"key":"e_1_3_2_2_2","doi-asserted-by":"crossref","first-page":"65","DOI":"10.1007\/978-3-319-66963-2_7","volume-title":"Applied Computer Sciences in Engineering","author":"Aguirre Santiago","year":"2017","unstructured":"Santiago Aguirre and Alejandro Rodriguez. 2017. Automation of a business process using robotic process automation (RPA): A case study. In Applied Computer Sciences in Engineering. Juan Carlos Figueroa-Garc\u00eda, Eduyn Ramiro L\u00f3pez-Santana, Jos\u00e9 Luis Villa-Ram\u00edrez, and Roberto Ferro-Escobar (Eds.), Springer International Publishing, Cham, 65\u201371."},{"unstructured":"Michael Ahn Anthony Brohan Noah Brown Yevgen Chebotar Omar Cortes Byron David Chelsea Finn Chuyuan Fu Keerthana Gopalakrishnan Karol Hausman et al. 2022. Do as I can not as I say: Grounding language in robotic affordances. arXiv:2204.01691. Retrieved from https:\/\/arxiv.org\/abs\/2204.01691","key":"e_1_3_2_3_2"},{"key":"e_1_3_2_4_2","doi-asserted-by":"crossref","first-page":"568","DOI":"10.1007\/978-3-642-34478-7_69","volume-title":"Neural Information Processing","author":"Al-Razgan Muna S.","year":"2012","unstructured":"Muna S. Al-Razgan, Hend S. Al-Khalifa, Mona D. Al-Shahrani, and Hessah H. AlAjmi. 2012. Touch-based mobile phone interface guidelines and design recommendations for elderly people: A survey of the literature. In Neural Information Processing. Tingwen Huang, Zhigang Zeng, Chuandong Li, and Chi Sing Leung (Eds.), Springer, Berlin, 568\u2013574."},{"doi-asserted-by":"publisher","key":"e_1_3_2_5_2","DOI":"10.1609\/aimag.v35i4.2513"},{"doi-asserted-by":"publisher","key":"e_1_3_2_6_2","DOI":"10.24963\/ijcai.2021\/235"},{"doi-asserted-by":"publisher","key":"e_1_3_2_7_2","DOI":"10.1145\/3290605.3300473"},{"doi-asserted-by":"publisher","key":"e_1_3_2_8_2","DOI":"10.1038\/s41592-019-0582-9"},{"doi-asserted-by":"publisher","key":"e_1_3_2_9_2","DOI":"10.1016\/0169-7552(87)90085-7"},{"doi-asserted-by":"publisher","key":"e_1_3_2_10_2","DOI":"10.5555\/1687878.1687892"},{"doi-asserted-by":"crossref","unstructured":"Andrea Burns Deniz Arsan Sanjna Agrawal Ranjitha Kumar Kate Saenko and Bryan A. Plummer. 2022. A dataset for interactive vision-language navigation with unknown command feasibility. arXiv:2202.02312. Retrieved from http:\/\/arxiv.org\/abs\/2202.02312","key":"e_1_3_2_11_2","DOI":"10.1007\/978-3-031-20074-8_18"},{"doi-asserted-by":"publisher","key":"e_1_3_2_12_2","DOI":"10.1145\/3334480.3382839"},{"doi-asserted-by":"publisher","key":"e_1_3_2_13_2","DOI":"10.1109\/CVPR52688.2022.01750"},{"doi-asserted-by":"publisher","key":"e_1_3_2_14_2","DOI":"10.1145\/3491102.3502073"},{"doi-asserted-by":"publisher","key":"e_1_3_2_15_2","DOI":"10.4236\/adr.2017.53013"},{"doi-asserted-by":"crossref","unstructured":"Yi Chen Rui Wang Haiyun Jiang Shuming Shi and Ruifeng Xu. 2023. Exploring the use of large language models for reference-free text quality evaluation: An empirical study. arXiv:2304.00723. Retrieved from https:\/\/arxiv.org\/abs\/2304.00723","key":"e_1_3_2_16_2","DOI":"10.18653\/v1\/2023.findings-ijcnlp.32"},{"issue":"1","key":"e_1_3_2_17_2","doi-asserted-by":"crossref","first-page":"11267","DOI":"10.1038\/s41598-020-68273-y","article-title":"Interactive machine learning for soybean seed and seedling quality classification","volume":"10","author":"de Medeiros Andr\u00e9 Dantas","year":"2020","unstructured":"Andr\u00e9 Dantas de Medeiros, Nayara Pereira Capobiango, Jose Maria da Silva, Laercio Junio da Silva, Clissia Barboza da Silva, and Denise Cunha Fernandes dos Santos Dias. 2020. Interactive machine learning for soybean seed and seedling quality classification. Scientific Reports 10, 1 (2020), 11267.","journal-title":"Scientific Reports"},{"doi-asserted-by":"publisher","key":"e_1_3_2_18_2","DOI":"10.1145\/3126594.3126651"},{"doi-asserted-by":"publisher","key":"e_1_3_2_19_2","DOI":"10.1016\/j.isci.2020.101656"},{"doi-asserted-by":"publisher","key":"e_1_3_2_20_2","DOI":"10.1145\/2556288.2556979"},{"doi-asserted-by":"publisher","key":"e_1_3_2_21_2","DOI":"10.1109\/ACCESS.2018.2831228"},{"unstructured":"Yilun Du Shuang Li Antonio Torralba Joshua B. Tenenbaum and Igor Mordatch. 2023. Improving factuality and reasoning in language models through multiagent debate. arXiv:2305.14325. Retrieved from https:\/\/arxiv.org\/abs\/2305.14325","key":"e_1_3_2_22_2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_23_2","DOI":"10.1145\/3185517"},{"doi-asserted-by":"publisher","key":"e_1_3_2_24_2","DOI":"10.1145\/604045.604056"},{"doi-asserted-by":"publisher","key":"e_1_3_2_25_2","DOI":"10.18653\/v1\/2022.acl-long.62"},{"doi-asserted-by":"publisher","key":"e_1_3_2_26_2","DOI":"10.1145\/3597503.3608137"},{"doi-asserted-by":"publisher","key":"e_1_3_2_27_2","DOI":"10.1145\/3290605.3300527"},{"doi-asserted-by":"publisher","key":"e_1_3_2_28_2","DOI":"10.1145\/3290605.3300439"},{"doi-asserted-by":"publisher","key":"e_1_3_2_29_2","DOI":"10.1609\/aaai.v35i7.16741"},{"doi-asserted-by":"publisher","key":"e_1_3_2_30_2","DOI":"10.1007\/s10257-022-00553-8"},{"unstructured":"Sirui Hong Xiawu Zheng Jonathan Chen Yuheng Cheng Jinlin Wang Ceyao Zhang Zili Wang Steven Ka Shing Yau Zijuan Lin Liyang Zhou et al. 2023. MetaGPT: Meta programming for multi-agent collaborative framework. arXiv:2308.00352. Retrieved from https:\/\/arxiv.org\/abs\/2308.00352","key":"e_1_3_2_31_2"},{"unstructured":"Wenyi Hong Weihan Wang Qingsong Lv Jiazheng Xu Wenmeng Yu Junhui Ji Yan Wang Zihan Wang Yuxuan Zhang Juanzi Li et al. 2023. CogAgent: A visual language model for GUI agents. arXiv:2312.08914. Retrieved from https:\/\/arxiv.org\/abs\/2312.08914","key":"e_1_3_2_32_2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_33_2","DOI":"10.1145\/3613904.3642074"},{"doi-asserted-by":"publisher","key":"e_1_3_2_34_2","DOI":"10.1145\/3610929"},{"issue":"6","key":"e_1_3_2_35_2","first-page":"72","article-title":"Mobile application and its global impact","volume":"10","author":"Islam Rashedul","year":"2010","unstructured":"Rashedul Islam, Rofiqul Islam, and Tohidul Mazumder. 2010. Mobile application and its global impact. International Journal of Engineering & Technology 10, 6 (2010), 72\u201378.","journal-title":"International Journal of Engineering & Technology"},{"doi-asserted-by":"publisher","key":"e_1_3_2_36_2","DOI":"10.1145\/281250.281253"},{"doi-asserted-by":"publisher","key":"e_1_3_2_37_2","DOI":"10.1145\/3571730"},{"doi-asserted-by":"publisher","key":"e_1_3_2_38_2","DOI":"10.1145\/3550321"},{"issue":"3","key":"e_1_3_2_39_2","first-page":"34","article-title":"Robotic process automation: Overview and opportunities","volume":"46","author":"Jovanovi\u0107 Stefan Z.","year":"2018","unstructured":"Stefan Z. Jovanovi\u0107, Jelena S. \u00d0uri\u0107, and Tatjana V. \u0160ibalija. 2018. Robotic process automation: Overview and opportunities. International Journal Advanced Quality 46, 3\u20134 (2018), 34\u201339.","journal-title":"International Journal Advanced Quality"},{"unstructured":"Kenton Lee Mandar Joshi Iulia Turc Hexiang Hu Fangyu Liu Julian Eisenschlos Urvashi Khandelwal Peter Shaw Ming-Wei Chang and Kristina Toutanova. 2022. Pix2Struct: Screenshot parsing as pretraining for visual language understanding. arXiv:2210.03347. Retrieved from http:\/\/arxiv.org\/abs\/2210.03347","key":"e_1_3_2_40_2"},{"key":"e_1_3_2_41_2","series-title":"Proceedings of Machine Learning Research","first-page":"18893","volume-title":"Proceedings of the 40th International Conference on Machine Learning","volume":"202","author":"Lee Kenton","year":"2023","unstructured":"Kenton Lee, Mandar Joshi, Iulia Raluca Turc, Hexiang Hu, Fangyu Liu, Julian Martin Eisenschlos, Urvashi Khandelwal, Peter Shaw, Ming-Wei Chang, and Kristina Toutanova. 2023. Pix2Struct: Screenshot parsing as pretraining for visual language understanding. In Proceedings of the 40th International Conference on Machine Learning. Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (Eds.), Proceedings of Machine Learning Research, Vol. 202, PMLR, 18893\u201318912. Retrieved from https:\/\/proceedings.mlr.press\/v202\/lee23g.html"},{"unstructured":"Gang Li and Yang Li. 2023. Spotlight: Mobile UI understanding using vision-language models with a focus. arXiv:2209.14927. Retrieved from https:\/\/arxiv.org\/abs\/2209.14927","key":"e_1_3_2_42_2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_43_2","DOI":"10.1109\/UIC-ATC-ScalCom-CBDCom-IoP.2015.50"},{"unstructured":"Tao Li Gang Li Jingjie Zheng Purple Wang and Yang Li. 2024. MUG: Interactive multimodal grounding on user interfaces. In Findings of the Association for Computational Linguistics (EACL \u201924). Yvette Graham and Matthew Purver (Eds.) Association for Computational Linguistics St. Julian\u2019s Malta 231\u2013251. Retrieved from https:\/\/aclanthology.org\/2024.findings-eacl.17","key":"e_1_3_2_44_2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_45_2","DOI":"10.1145\/3025453.3025483"},{"doi-asserted-by":"publisher","key":"e_1_3_2_46_2","DOI":"10.1109\/VLHCC.2018.8506506"},{"doi-asserted-by":"publisher","key":"e_1_3_2_47_2","DOI":"10.1145\/3411764.3445049"},{"doi-asserted-by":"publisher","key":"e_1_3_2_48_2","DOI":"10.1145\/3210240.3210339"},{"unstructured":"Wei Li. 2021. Learning UI navigation through demonstrations composed of macro actions. arXiv:2110.08653. Retrieved from http:\/\/arxiv.org\/abs\/2110.08653","key":"e_1_3_2_49_2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_50_2","DOI":"10.18653\/v1\/2020.acl-main.729"},{"unstructured":"Yang Li Gang Li Luheng He Jingjie Zheng Hong Li and Zhiwei Guan. 2020. Widget captioning: Generating natural language description for mobile user interface elements. arXiv:2010.04295. Retrieved from http:\/\/arxiv.org\/abs\/2010.04295","key":"e_1_3_2_51_2"},{"unstructured":"Zhangheng Li Keen You Haotian Zhang Di Feng Harsh Agrawal Xiujun Li Mohana Prasad Sathya Moorthy Jeff Nichols Yinfei Yang and Zhe Gan. 2024. Ferret-UI 2: Mastering universal user interface understanding across platforms. arXiv:2410.18967. Retrieved from https:\/\/arxiv.org\/abs\/2410.18967","key":"e_1_3_2_52_2"},{"unstructured":"Tian Liang Zhiwei He Wenxiang Jiao Xing Wang Yan Wang Rui Wang Yujiu Yang Zhaopeng Tu and Shuming Shi. 2023. Encouraging divergent thinking in large language models through multi-agent debate. arXiv:2305.19118. Retrieved from https:\/\/arxiv.org\/abs\/2305.19118","key":"e_1_3_2_53_2"},{"key":"e_1_3_2_54_2","first-page":"135","article-title":"Comparing task models for user interface design","volume":"6","author":"Limbourg Quentin","year":"2004","unstructured":"Quentin Limbourg and Jean Vanderdonckt. 2004. Comparing task models for user interface design. The Handbook of Task Analysis for Human-Computer Interaction 6 (2004), 135\u2013154.","journal-title":"The Handbook of Task Analysis for Human-Computer Interaction"},{"doi-asserted-by":"publisher","key":"e_1_3_2_55_2","DOI":"10.1145\/3242587.3242650"},{"doi-asserted-by":"publisher","key":"e_1_3_2_56_2","DOI":"10.1007\/s00521-020-04805-x"},{"unstructured":"Norman Di Palo Arunkumar Byravan Leonard Hasenclever Markus Wulfmeier Nicolas Heess and Martin Riedmiller. 2023. Towards a unified agent with foundation models. arXiv:2307.09668. Retrieved from https:\/\/arxiv.org\/abs\/2307.09668","key":"e_1_3_2_57_2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_58_2","DOI":"10.1145\/3491102.3517459"},{"doi-asserted-by":"publisher","key":"e_1_3_2_59_2","DOI":"10.1007\/s10458-005-2631-2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_60_2","DOI":"10.1016\/j.displa.2008.07.001"},{"doi-asserted-by":"crossref","unstructured":"Panupong Pasupat Tian-Shun Jiang Evan Zheran Liu Kelvin Guu and Percy Liang. 2018. Mapping natural language commands to web elements. arXiv:1808.09132. Retrieved from https:\/\/arxiv.org\/abs\/1808.09132","key":"e_1_3_2_61_2","DOI":"10.18653\/v1\/D18-1540"},{"doi-asserted-by":"publisher","key":"e_1_3_2_62_2","DOI":"10.1007\/978-0-387-35175-9_58"},{"doi-asserted-by":"publisher","key":"e_1_3_2_63_2","DOI":"10.1145\/2047196.2047213"},{"doi-asserted-by":"publisher","key":"e_1_3_2_64_2","DOI":"10.1016\/j.procs.2021.01.104"},{"doi-asserted-by":"publisher","key":"e_1_3_2_65_2","DOI":"10.1002\/0471741442.ch6"},{"unstructured":"Timo Schick Jane Dwivedi-Yu Roberto Dess\u00ec Roberta Raileanu Maria Lomeli Luke Zettlemoyer Nicola Cancedda and Thomas Scialom. 2023. Toolformer: Language models can teach themselves to use tools. arXiv:2302.04761. Retrieved from https:\/\/arxiv.org\/abs\/2302.04761","key":"e_1_3_2_66_2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_67_2","DOI":"10.1148\/radiol.230163"},{"doi-asserted-by":"publisher","key":"e_1_3_2_68_2","DOI":"10.1109\/ICRA48891.2023.10161317"},{"doi-asserted-by":"publisher","key":"e_1_3_2_69_2","DOI":"10.1145\/3654777.3676386"},{"doi-asserted-by":"publisher","key":"e_1_3_2_70_2","DOI":"10.1109\/VAST47406.2019.8986917"},{"doi-asserted-by":"publisher","key":"e_1_3_2_71_2","DOI":"10.1016\/j.apergo.2005.06.003"},{"doi-asserted-by":"publisher","key":"e_1_3_2_72_2","DOI":"10.1145\/3613904.3642777"},{"unstructured":"Yashar Talebirad and Amirhossein Nadiri. 2023. Multi-agent collaboration: Harnessing the power of intelligent LLM agents. arXiv:2306.03314. Retrieved from https:\/\/arxiv.org\/abs\/2306.03314","key":"e_1_3_2_73_2"},{"unstructured":"Sagar Gubbi Venkatesh Partha Talukdar and Srini Narayanan. 2022. UGIF: UI grounded instruction following. arXiv:2211.07615. Retrieved from http:\/\/arxiv.org\/abs\/2211.07615","key":"e_1_3_2_74_2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_75_2","DOI":"10.1145\/3654777.3676356"},{"doi-asserted-by":"publisher","key":"e_1_3_2_76_2","DOI":"10.1145\/3581998"},{"doi-asserted-by":"publisher","key":"e_1_3_2_77_2","DOI":"10.1145\/3544548.3580895"},{"doi-asserted-by":"publisher","key":"e_1_3_2_78_2","DOI":"10.1145\/3472749.3474765"},{"doi-asserted-by":"publisher","key":"e_1_3_2_79_2","DOI":"10.1145\/2556288.2557407"},{"unstructured":"Junyang Wang Haiyang Xu Jiabo Ye Ming Yan Weizhou Shen Ji Zhang Fei Huang and Jitao Sang. 2024. Mobile-agent: Autonomous multi-modal mobile device agent with visual perception. arXiv:2401.16158. Retrieved from https:\/\/arxiv.org\/abs\/2401.16158","key":"e_1_3_2_80_2"},{"unstructured":"Jimmy Wei Kurt Shuster Arthur Szlam Jason Weston Jack Urbanek and Mojtaba Komeili. 2023. Multi-party chat: Conversational agents in group settings with humans and models. arXiv:2304.13835. Retrieved from https:\/\/arxiv.org\/abs\/2304.13835","key":"e_1_3_2_81_2"},{"doi-asserted-by":"crossref","unstructured":"Hao Wen Yuanchun Li Guohong Liu Shanhui Zhao Tao Yu Toby Jia-Jun Li Shiqi Jiang Yunhao Liu Yaqin Zhang and Yunxin Liu. 2024. AutoDroid: LLM-powered task automation in Android. arXiv:2308.15272. Retrieved from https:\/\/arxiv.org\/abs\/2308.15272","key":"e_1_3_2_82_2","DOI":"10.1145\/3636534.3649379"},{"doi-asserted-by":"publisher","key":"e_1_3_2_83_2","DOI":"10.1145\/371920.371974"},{"doi-asserted-by":"publisher","key":"e_1_3_2_84_2","DOI":"10.1109\/EDOC49727.2020.00021"},{"doi-asserted-by":"publisher","key":"e_1_3_2_85_2","DOI":"10.1145\/3293882.3330551"},{"doi-asserted-by":"publisher","key":"e_1_3_2_86_2","DOI":"10.1145\/3586183.3606824"},{"doi-asserted-by":"publisher","key":"e_1_3_2_87_2","DOI":"10.1145\/3368089.3417940"},{"unstructured":"Shunyu Yao Jeffrey Zhao Dian Yu Nan Du Izhak Shafran Karthik Narasimhan and Yuan Cao. 2022. ReAct: Synergizing reasoning and acting in language models. arXiv:2210.03629. Retrieved from https:\/\/arxiv.org\/abs\/2210.03629","key":"e_1_3_2_88_2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_89_2","DOI":"10.1007\/978-3-031-73039-9_14"},{"unstructured":"Hongxin Zhang Weihua Du Jiaming Shan Qinhong Zhou Yilun Du Joshua B. Tenenbaum Tianmin Shu and Chuang Gan. 2023. Building cooperative embodied agents modularly with large language models. arXiv:2307.02485. Retrieved from https:\/\/arxiv.org\/abs\/2307.02485","key":"e_1_3_2_90_2"},{"unstructured":"Zhizheng Zhang Xiaoyi Zhang Wenxuan Xie and Yan Lu. 2023. Responsible task automation: Empowering large language models as responsible task automators. arXiv:2306.01242. Retrieved from https:\/\/arxiv.org\/abs\/2306.01242","key":"e_1_3_2_91_2"},{"unstructured":"Andrew Zhao Daniel Huang Quentin Xu Matthieu Lin Yong-Jin Liu and Gao Huang. 2023. ExpeL: LLM agents are experiential learners. arXiv:2308.10144. Retrieved from https:\/\/arxiv.org\/abs\/2308.10144","key":"e_1_3_2_92_2"},{"doi-asserted-by":"publisher","key":"e_1_3_2_93_2","DOI":"10.1145\/2702123.2702588"},{"doi-asserted-by":"publisher","key":"e_1_3_2_94_2","DOI":"10.1145\/3472749.3474812"}],"container-title":["ACM Transactions on Computer-Human Interaction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3716132","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,14]],"date-time":"2025-06-14T16:15:25Z","timestamp":1749917725000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3716132"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,14]]},"references-count":93,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,6,30]]}},"alternative-id":["10.1145\/3716132"],"URL":"https:\/\/doi.org\/10.1145\/3716132","relation":{},"ISSN":["1073-0516","1557-7325"],"issn-type":[{"type":"print","value":"1073-0516"},{"type":"electronic","value":"1557-7325"}],"subject":[],"published":{"date-parts":[[2025,6,14]]},"assertion":[{"value":"2024-02-09","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-01-07","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-06-14","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}