{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,4]],"date-time":"2026-05-04T10:11:09Z","timestamp":1777889469158,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":33,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,11,25]],"date-time":"2023-11-25T00:00:00Z","timestamp":1700870400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,11,27]]},"DOI":"10.1145\/3604237.3626908","type":"proceedings-article","created":{"date-parts":[[2023,11,25]],"date-time":"2023-11-25T18:09:47Z","timestamp":1700935787000},"page":"73-81","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":36,"title":["FlowMind: Automatic Workflow Generation with LLMs"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4383-3590","authenticated-orcid":false,"given":"Zhen","family":"Zeng","sequence":"first","affiliation":[{"name":"J.P. Morgan AI Research, US"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5516-262X","authenticated-orcid":false,"given":"William","family":"Watson","sequence":"additional","affiliation":[{"name":"J.P. Morgan AI Research, US"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-9007-3255","authenticated-orcid":false,"given":"Nicole","family":"Cho","sequence":"additional","affiliation":[{"name":"J.P. Morgan AI Research, US"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6793-8202","authenticated-orcid":false,"given":"Saba","family":"Rahimi","sequence":"additional","affiliation":[{"name":"J.P. Morgan AI Research, US"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-5905-4189","authenticated-orcid":false,"given":"Shayleen","family":"Reynolds","sequence":"additional","affiliation":[{"name":"J.P. Morgan AI Research, US"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5148-2033","authenticated-orcid":false,"given":"Tucker","family":"Balch","sequence":"additional","affiliation":[{"name":"J.P. Morgan AI Research, US"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6738-238X","authenticated-orcid":false,"given":"Manuela","family":"Veloso","sequence":"additional","affiliation":[{"name":"J.P. Morgan AI Research, US"}]}],"member":"320","published-online":{"date-parts":[[2023,11,25]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"1984. Edgar. https:\/\/www.sec.gov\/edgar."},{"key":"e_1_3_2_1_2_1","unstructured":"2022. LangChain. https:\/\/github.com\/langchain-ai\/langchain."},{"key":"e_1_3_2_1_3_1","unstructured":"2022. OpenAI API. https:\/\/platform.openai.com\/docs\/guides\/embeddings."},{"key":"e_1_3_2_1_4_1","unstructured":"2023. AutoGPT. https:\/\/github.com\/Significant-Gravitas\/Auto-GPT."},{"key":"e_1_3_2_1_5_1","unstructured":"2023. Transformer Agent. https:\/\/huggingface.co\/docs\/transformers\/main_classes\/agent."},{"key":"e_1_3_2_1_6_1","volume-title":"Do as i can, not as i say: Grounding language in robotic affordances. arXiv preprint arXiv:2204.01691","author":"Ahn Michael","year":"2022","unstructured":"Michael Ahn, Anthony Brohan, Noah Brown, Yevgen Chebotar, Omar Cortes, Byron David, Chelsea Finn, Chuyuan Fu, Keerthana Gopalakrishnan, Karol Hausman, 2022. Do as i can, not as i say: Grounding language in robotic affordances. arXiv preprint arXiv:2204.01691 (2022)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.5281\/zenodo.5584996"},{"key":"e_1_3_2_1_8_1","volume-title":"PAL: Program-aided Language Models. ArXiv abs\/2211.10435","author":"Gao Luyu","year":"2022","unstructured":"Luyu Gao, Aman Madaan, Shuyan Zhou, Uri Alon, Pengfei Liu, Yiming Yang, Jamie Callan, and Graham Neubig. 2022. PAL: Program-aided Language Models. ArXiv abs\/2211.10435 (2022)."},{"key":"e_1_3_2_1_9_1","volume-title":"A Consolidated Framework for Implementing Robotic Process Automation Projects","author":"Herm Lukas-Valentin","unstructured":"Lukas-Valentin Herm, Christian Janiesch, Alexander Helm, Florian Imgrund, Kevin Fuchs, Adrian Hofmann, and Axel Winkelmann. 2020. A Consolidated Framework for Implementing Robotic Process Automation Projects. In Business Process Management, Dirk Fahland, Chiara Ghidini, J\u00f6rg Becker, and Marlon Dumas (Eds.). Springer International Publishing, Cham, 471\u2013488."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1007\/s12525-019-00365-8"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/COMPSAC.2019.10215"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48891.2023.10160591"},{"key":"e_1_3_2_1_13_1","volume-title":"Is your code generated by chatgpt really correct? rigorous evaluation of large language models for code generation. arXiv preprint arXiv:2305.01210","author":"Liu Jiawei","year":"2023","unstructured":"Jiawei Liu, Chunqiu\u00a0Steven Xia, Yuyao Wang, and Lingming Zhang. 2023. Is your code generated by chatgpt really correct? rigorous evaluation of large language models for code generation. arXiv preprint arXiv:2305.01210 (2023)."},{"key":"e_1_3_2_1_14_1","volume-title":"FinGPT: Democratizing Internet-scale Data for Financial Large Language Models. arXiv preprint arXiv:2307.10485","author":"Liu Xiao-Yang","year":"2023","unstructured":"Xiao-Yang Liu, Guoxuan Wang, and Daochen Zha. 2023. FinGPT: Democratizing Internet-scale Data for Financial Large Language Models. arXiv preprint arXiv:2307.10485 (2023)."},{"key":"e_1_3_2_1_15_1","unstructured":"Reiichiro Nakano Jacob Hilton Suchir Balaji Jeff Wu Long Ouyang Christina Kim Christopher Hesse Shantanu Jain Vineet Kosaraju William Saunders Xu Jiang Karl Cobbe Tyna Eloundou Gretchen Krueger Kevin Button Matthew Knight Benjamin Chess and John Schulman. 2022. WebGPT: Browser-assisted question-answering with human feedback. arxiv:2112.09332\u00a0[cs.CL]"},{"key":"e_1_3_2_1_16_1","volume-title":"Codegen: An open large language model for code with multi-turn program synthesis. arXiv preprint arXiv:2203.13474","author":"Nijkamp Erik","year":"2022","unstructured":"Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, and Caiming Xiong. 2022. Codegen: An open large language model for code with multi-turn program synthesis. arXiv preprint arXiv:2203.13474 (2022)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-28238-6_44"},{"key":"e_1_3_2_1_18_1","volume-title":"Synchromesh: Reliable code generation from pre-trained language models. arXiv preprint arXiv:2201.11227","author":"Poesia Gabriel","year":"2022","unstructured":"Gabriel Poesia, Oleksandr Polozov, Vu Le, Ashish Tiwari, Gustavo Soares, Christopher Meek, and Sumit Gulwani. 2022. Synchromesh: Reliable code generation from pre-trained language models. arXiv preprint arXiv:2201.11227 (2022)."},{"key":"e_1_3_2_1_19_1","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans Ilya Sutskever 2018. Improving language understanding by generative pre-training. (2018)."},{"key":"e_1_3_2_1_20_1","unstructured":"Alec Radford Jeff Wu Rewon Child David Luan Dario Amodei and Ilya Sutskever. 2019. Language Models are Unsupervised Multitask Learners. (2019)."},{"key":"e_1_3_2_1_21_1","volume-title":"In-context retrieval-augmented language models. arXiv preprint arXiv:2302.00083","author":"Ram Ori","year":"2023","unstructured":"Ori Ram, Yoav Levine, Itay Dalmedigos, Dor Muhlgay, Amnon Shashua, Kevin Leyton-Brown, and Yoav Shoham. 2023. In-context retrieval-augmented language models. arXiv preprint arXiv:2302.00083 (2023)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3275116.3275129"},{"key":"e_1_3_2_1_23_1","volume-title":"Learning to retrieve prompts for in-context learning. arXiv preprint arXiv:2112.08633","author":"Rubin Ohad","year":"2021","unstructured":"Ohad Rubin, Jonathan Herzig, and Jonathan Berant. 2021. Learning to retrieve prompts for in-context learning. arXiv preprint arXiv:2112.08633 (2021)."},{"key":"e_1_3_2_1_24_1","volume-title":"Toolformer: Language Models Can Teach Themselves to Use Tools.","author":"Schick Timo","year":"2023","unstructured":"Timo Schick, Jane Dwivedi-Yu, Roberto Dessi, Roberta Raileanu, Maria Lomeli, Luke Zettlemoyer, Nicola Cancedda, and Thomas Scialom. 2023. Toolformer: Language Models Can Teach Themselves to Use Tools."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"crossref","unstructured":"Sanjay Subramanian Medhini Narasimhan Kushal Khangaonkar Kevin Yang Arsha Nagrani Cordelia Schmid Andy Zeng Trevor Darrell and Dan Klein. 2023. Modular Visual Question Answering via Code Generation. arxiv:2306.05392\u00a0[cs.CL]","DOI":"10.18653\/v1\/2023.acl-short.65"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.compind.2019.103162"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491101.3519665"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1007\/s12599-018-0542-4"},{"key":"e_1_3_2_1_29_1","first-page":"20","article-title":"Chatgpt for robotics: Design principles and model abilities","volume":"2","author":"Vemprala Sai","year":"2023","unstructured":"Sai Vemprala, Rogerio Bonatti, Arthur Bucker, and Ashish Kapoor. 2023. Chatgpt for robotics: Design principles and model abilities. Microsoft Auton. Syst. Robot. Res 2 (2023), 20.","journal-title":"Microsoft Auton. Syst. Robot. Res"},{"key":"e_1_3_2_1_30_1","first-page":"71","article-title":"Robotic process automation in banking industry: a case study on Deutsche Bank","volume":"5","author":"Villar Alice\u00a0Saldanha","year":"2021","unstructured":"Alice\u00a0Saldanha Villar and Nawaz Khan. 2021. Robotic process automation in banking industry: a case study on Deutsche Bank. Journal of Banking and Financial Technology 5, 1 (2021), 71\u201386.","journal-title":"Journal of Banking and Financial Technology"},{"key":"e_1_3_2_1_31_1","unstructured":"Shijie Wu Ozan Irsoy Steven Lu Vadim Dabravolski Mark Dredze Sebastian Gehrmann Prabhanjan Kambadur David Rosenberg and Gideon Mann. 2023. BloombergGPT: A Large Language Model for Finance. arxiv:2303.17564\u00a0[cs.LG]"},{"key":"e_1_3_2_1_32_1","volume-title":"FinGPT: Open-Source Financial Large Language Models. arXiv preprint arXiv:2306.06031","author":"Yang Hongyang","year":"2023","unstructured":"Hongyang Yang, Xiao-Yang Liu, and Christina\u00a0Dan Wang. 2023. FinGPT: Open-Source Financial Large Language Models. arXiv preprint arXiv:2306.06031 (2023)."},{"key":"e_1_3_2_1_33_1","volume-title":"Instruct-FinGPT: Financial Sentiment Analysis by Instruction Tuning of General-Purpose Large Language Models. arXiv preprint arXiv:2306.12659","author":"Zhang Boyu","year":"2023","unstructured":"Boyu Zhang, Hongyang Yang, and Xiao-Yang Liu. 2023. Instruct-FinGPT: Financial Sentiment Analysis by Instruction Tuning of General-Purpose Large Language Models. arXiv preprint arXiv:2306.12659 (2023)."}],"event":{"name":"ICAIF '23: 4th ACM International Conference on AI in Finance","location":"Brooklyn NY USA","acronym":"ICAIF '23"},"container-title":["4th ACM International Conference on AI in Finance"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3604237.3626908","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3604237.3626908","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T17:39:34Z","timestamp":1755884374000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3604237.3626908"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,25]]},"references-count":33,"alternative-id":["10.1145\/3604237.3626908","10.1145\/3604237"],"URL":"https:\/\/doi.org\/10.1145\/3604237.3626908","relation":{},"subject":[],"published":{"date-parts":[[2023,11,25]]},"assertion":[{"value":"2023-11-25","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}