{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T17:16:07Z","timestamp":1775841367130,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":38,"publisher":"ACM","funder":[{"name":"the National Key R&D Program of China","award":["Grant No.2023YFF0725001"],"award-info":[{"award-number":["Grant No.2023YFF0725001"]}]},{"name":"the National Natural Science Foundation of China","award":["Grant No.92370204"],"award-info":[{"award-number":["Grant No.92370204"]}]},{"name":"the Guangdong Basic and Applied Basic Research Foundation","award":["Grant No.2023B1515120057"],"award-info":[{"award-number":["Grant No.2023B1515120057"]}]},{"name":"the Key-Area Special Project of Guangdong Provincial Ordinary Universities","award":["2024ZDZX1007"],"award-info":[{"award-number":["2024ZDZX1007"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2026,4,13]]},"DOI":"10.1145\/3774904.3792817","type":"proceedings-article","created":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T21:54:39Z","timestamp":1775771679000},"page":"7845-7856","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["ARADD: An Automatic Real-World API Discovery and Deployment Framework for AI Guide Service in Baidu Map"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-6441-2390","authenticated-orcid":false,"given":"Fuling","family":"Wang","sequence":"first","affiliation":[{"name":"Hong Kong University of Science and Technology(Guangzhou), Guangzhou, Guangdong, China and Baidu Inc., Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0894-9651","authenticated-orcid":false,"given":"Le","family":"Zhang","sequence":"additional","affiliation":[{"name":"Baidu Inc., Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2677-7021","authenticated-orcid":false,"given":"Jingbo","family":"Zhou","sequence":"additional","affiliation":[{"name":"Baidu Inc., Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1542-6149","authenticated-orcid":false,"given":"Jindong","family":"Han","sequence":"additional","affiliation":[{"name":"Shandong University, Jinan, Shandong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4763-6060","authenticated-orcid":false,"given":"Ying","family":"Sun","sequence":"additional","affiliation":[{"name":"Hong Kong University of Science and Technology (Guangzhou), Guangzhou, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5354-8630","authenticated-orcid":false,"given":"Chuan","family":"Qin","sequence":"additional","affiliation":[{"name":"Computer Network Information Center, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4570-643X","authenticated-orcid":false,"given":"Hengshu","family":"Zhu","sequence":"additional","affiliation":[{"name":"Computer Network Information Center, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6016-6465","authenticated-orcid":false,"given":"Hui","family":"Xiong","sequence":"additional","affiliation":[{"name":"Hong Kong University of Science and Technology (Guangzhou), Guangzhou, Guangdong, China and Hong Kong University of Science and Technology Hong Kong SAR, Hong Kong, China"}]}],"member":"320","published-online":{"date-parts":[[2026,4,12]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"LAMP: A Language Model on the Map. arXiv preprint arXiv:2403.09059","author":"Balsebre Pasquale","year":"2024","unstructured":"Pasquale Balsebre, Weiming Huang, and Gao Cong. 2024. LAMP: A Language Model on the Map. arXiv preprint arXiv:2403.09059 (2024)."},{"key":"e_1_3_2_1_2_1","volume-title":"Large language models as tool makers. arXiv preprint arXiv:2305.17126","author":"Cai Tianle","year":"2023","unstructured":"Tianle Cai, Xuezhi Wang, Tengyu Ma, Xinyun Chen, and Denny Zhou. 2023a. Large language models as tool makers. arXiv preprint arXiv:2305.17126 (2023)."},{"key":"e_1_3_2_1_3_1","volume-title":"Large language models as tool makers. arXiv preprint arXiv:2305.17126","author":"Cai Tianle","year":"2023","unstructured":"Tianle Cai, Xuezhi Wang, Tengyu Ma, Xinyun Chen, and Denny Zhou. 2023b. Large language models as tool makers. arXiv preprint arXiv:2305.17126 (2023)."},{"key":"e_1_3_2_1_4_1","volume-title":"Efficient intent detection with dual sentence encoders. arXiv preprint arXiv:2003.04807","author":"Casanueva I","year":"2020","unstructured":"I nigo Casanueva, Tadas Tem\u010dinas, Daniela Gerz, Matthew Henderson, and Ivan Vuli\u0107. 2020. Efficient intent detection with dual sentence encoders. arXiv preprint arXiv:2003.04807 (2020)."},{"key":"e_1_3_2_1_5_1","volume-title":"Xing","author":"Chiang Wei-Lin","year":"2023","unstructured":"Wei-Lin Chiang, Zhuohan Li, Zi Lin, Ying Sheng, Zhanghao Wu, Hao Zhang, Lianmin Zheng, Siyuan Zhuang, Yonghao Zhuang, Joseph E. Gonzalez, Ion Stoica, and Eric P. Xing. 2023. Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality. https:\/\/lmsys.org\/blog\/2023-03-30-vicuna\/"},{"key":"e_1_3_2_1_6_1","first-page":"1854","article-title":"Clustering novel intents in a conversational interaction system with semantic parsing","author":"Hakkani-T\u00fcr Dilek","year":"2015","unstructured":"Dilek Hakkani-T\u00fcr, Yun-Cheng Ju, Geoffrey Zweig, and G\u00f6khan T\u00fcr. 2015. Clustering novel intents in a conversational interaction system with semantic parsing. In INTERSPEECH. 1854-1858.","journal-title":"INTERSPEECH."},{"key":"e_1_3_2_1_7_1","volume-title":"UrbanLLM: Autonomous Urban Activity Planning and Management with Large Language Models. arXiv preprint arXiv:2406.12360","author":"Jiang Yue","year":"2024","unstructured":"Yue Jiang, Qin Chao, Yile Chen, Xiucheng Li, Shuai Liu, and Gao Cong. 2024. UrbanLLM: Autonomous Urban Activity Planning and Management with Large Language Models. arXiv preprint arXiv:2406.12360 (2024)."},{"key":"e_1_3_2_1_8_1","unstructured":"Mingyu Jin Qinkai Yu Jingyuan Huang Qingcheng Zeng Zhenting Wang Wenyue Hua Haiyan Zhao Kai Mei Yanda Meng Kaize Ding et al. 2024. Exploring Concept Depth: How Large Language Models Acquire Knowledge and Concept at Different Layers? arXiv preprint arXiv:2404.07066 (2024)."},{"key":"e_1_3_2_1_9_1","volume-title":"Tan Yong Keat, and Fu Bin","author":"Junhua Liu","year":"2024","unstructured":"Liu Junhua, Tan Yong Keat, and Fu Bin. 2024. LARA: Linguistic-Adaptive Retrieval-Augmented LLMs for Multi-Turn Intent Classification. arXiv preprint arXiv:2403.16504 (2024)."},{"key":"e_1_3_2_1_10_1","volume-title":"Auto-Intent: Automated Intent Discovery and Self-Exploration for Large Language Model Web Agents. arXiv preprint arXiv:2410.22552","author":"Kim Jaekyeom","year":"2024","unstructured":"Jaekyeom Kim, Dong-Ki Kim, Lajanugen Logeswaran, Sungryull Sohn, and Honglak Lee. 2024. Auto-Intent: Automated Intent Discovery and Self-Exploration for Large Language Model Web Agents. arXiv preprint arXiv:2410.22552 (2024)."},{"key":"e_1_3_2_1_11_1","volume-title":"Api-bank: A comprehensive benchmark for tool-augmented llms. arXiv preprint arXiv:2304.08244","author":"Li Minghao","year":"2023","unstructured":"Minghao Li, Yingxiu Zhao, Bowen Yu, Feifan Song, Hangyu Li, Haiyang Yu, Zhoujun Li, Fei Huang, and Yongbin Li. 2023. Api-bank: A comprehensive benchmark for tool-augmented llms. arXiv preprint arXiv:2304.08244 (2023)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6353"},{"key":"e_1_3_2_1_13_1","unstructured":"Aixin Liu Bei Feng Bin Wang Bingxuan Wang Bo Liu Chenggang Zhao Chengqi Dengr Chong Ruan Damai Dai Daya Guo et al. 2024a. Deepseek-v2: A strong economical and efficient mixture-of-experts language model. arXiv preprint arXiv:2405.04434 (2024)."},{"key":"e_1_3_2_1_14_1","unstructured":"Aixin Liu Bei Feng Bing Xue Bingxuan Wang Bochao Wu Chengda Lu Chenggang Zhao Chengqi Deng Chenyu Zhang Chong Ruan et al. 2024b. Deepseek-v3 technical report. arXiv preprint arXiv:2412.19437 (2024)."},{"key":"e_1_3_2_1_15_1","volume-title":"ToolNet: Connecting large language models with massive tools via tool graph. arXiv preprint arXiv:2403.00839","author":"Liu Xukun","year":"2024","unstructured":"Xukun Liu, Zhiyuan Peng, Xiaoyuan Yi, Xing Xie, Lirong Xiang, Yuchen Liu, and Dongkuan Xu. 2024c. ToolNet: Connecting large language models with massive tools via tool graph. arXiv preprint arXiv:2403.00839 (2024)."},{"key":"e_1_3_2_1_16_1","volume-title":"Dawei Gao, Yaliang Li, and Ji-Rong Wen.","author":"Liu Zikang","year":"2024","unstructured":"Zikang Liu, Kun Zhou, Wayne Xin Zhao, Dawei Gao, Yaliang Li, and Ji-Rong Wen. 2024d. Less is More: Data Value Estimation for Visual Instruction Tuning. arXiv preprint arXiv:2403.09559 (2024)."},{"key":"e_1_3_2_1_17_1","volume-title":"International Conference on Machine Learning. PMLR, 22631-22648","author":"Longpre Shayne","year":"2023","unstructured":"Shayne Longpre, Le Hou, Tu Vu, Albert Webson, Hyung Won Chung, Yi Tay, Denny Zhou, Quoc V Le, Barret Zoph, Jason Wei, et al., 2023. The flan collection: Designing data and methods for effective instruction tuning. In International Conference on Machine Learning. PMLR, 22631-22648."},{"key":"e_1_3_2_1_18_1","unstructured":"Haoyu Lu Wen Liu Bo Zhang Bingxuan Wang Kai Dong Bo Liu Jingxiang Sun Tongzheng Ren Zhuoshu Li Hao Yang et al. 2024. Deepseek-vl: towards real-world vision-language understanding. arXiv preprint arXiv:2403.05525 (2024)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"crossref","unstructured":"Long Ouyang Jeffrey Wu Xu Jiang Diogo Almeida Carroll Wainwright Pamela Mishkin Chong Zhang Sandhini Agarwal Katarina Slama Alex Ray et al. 2022. Training language models to follow instructions with human feedback. Advances in neural information processing systems Vol. 35 (2022) 27730-27744.","DOI":"10.52202\/068431-2011"},{"key":"e_1_3_2_1_20_1","volume-title":"Gorilla: Large language model connected with massive apis. arXiv preprint arXiv:2305.15334","author":"Patil Shishir G","year":"2023","unstructured":"Shishir G Patil, Tianjun Zhang, Xin Wang, and Joseph E Gonzalez. 2023. Gorilla: Large language model connected with massive apis. arXiv preprint arXiv:2305.15334 (2023)."},{"key":"e_1_3_2_1_21_1","volume-title":"Instruction tuning with gpt-4. arXiv preprint arXiv:2304.03277","author":"Peng Baolin","year":"2023","unstructured":"Baolin Peng, Chunyuan Li, Pengcheng He, Michel Galley, and Jianfeng Gao. 2023. Instruction tuning with gpt-4. arXiv preprint arXiv:2304.03277 (2023)."},{"key":"e_1_3_2_1_22_1","first-page":"19920","article-title":"Estimating training data influence by tracing gradient descent","volume":"33","author":"Pruthi Garima","year":"2020","unstructured":"Garima Pruthi, Frederick Liu, Satyen Kale, and Mukund Sundararajan. 2020. Estimating training data influence by tracing gradient descent. Advances in Neural Information Processing Systems, Vol. 33 (2020), 19920-19930.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_23_1","volume-title":"Creator: Tool creation for disentangling abstract and concrete reasoning of large language models. arXiv preprint arXiv:2305.14318","author":"Qian Cheng","year":"2023","unstructured":"Cheng Qian, Chi Han, Yi R Fung, Yujia Qin, Zhiyuan Liu, and Heng Ji. 2023. Creator: Tool creation for disentangling abstract and concrete reasoning of large language models. arXiv preprint arXiv:2305.14318 (2023)."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3704435"},{"key":"e_1_3_2_1_25_1","volume-title":"Toolllm: Facilitating large language models to master 16000 real-world apis. arXiv preprint arXiv:2307.16789","author":"Qin Yujia","year":"2023","unstructured":"Yujia Qin, Shihao Liang, Yining Ye, Kunlun Zhu, Lan Yan, Yaxi Lu, Yankai Lin, Xin Cong, Xiangru Tang, Bill Qian, et al., 2023. Toolllm: Facilitating large language models to master 16000 real-world apis. arXiv preprint arXiv:2307.16789 (2023)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1072"},{"key":"e_1_3_2_1_27_1","volume-title":"Large Language Models Meet Open-World Intent Discovery and Recognition: An Evaluation of ChatGPT. arXiv preprint arXiv:2310","author":"Song Xiaoshuai","year":"2023","unstructured":"Xiaoshuai Song, Keqing He, Pei Wang, Guanting Dong, Yutao Mou, Jingang Wang, Yunsen Xian, Xunliang Cai, and Weiran Xu. 2023. Large Language Models Meet Open-World Intent Discovery and Recognition: An Evaluation of ChatGPT. arXiv preprint arXiv:2310.10176 (2023)."},{"key":"e_1_3_2_1_28_1","volume-title":"Alpaca: A strong, replicable instruction-following model","author":"Taori Rohan","year":"2023","unstructured":"Rohan Taori, Ishaan Gulrajani, Tianyi Zhang, Yann Dubois, Xuechen Li, Carlos Guestrin, Percy Liang, and Tatsunori B Hashimoto. 2023. Alpaca: A strong, replicable instruction-following model. Stanford Center for Research on Foundation Models. https:\/\/crfm. stanford. edu\/2023\/03\/13\/alpaca. html, Vol. 3, 6 (2023), 7."},{"key":"e_1_3_2_1_29_1","volume-title":"Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, et al., 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_1_30_1","first-page":"74764","article-title":"How far can camels go? exploring the state of instruction tuning on open resources","volume":"36","author":"Wang Yizhong","year":"2023","unstructured":"Yizhong Wang, Hamish Ivison, Pradeep Dasigi, Jack Hessel, Tushar Khot, Khyathi Chandu, David Wadden, Kelsey MacMillan, Noah A Smith, Iz Beltagy, et al., 2023. How far can camels go? exploring the state of instruction tuning on open resources. Advances in Neural Information Processing Systems, Vol. 36 (2023), 74764-74786.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_31_1","volume-title":"Self-instruct: Aligning language models with self-generated instructions. arXiv preprint arXiv:2212.10560","author":"Wang Yizhong","year":"2022","unstructured":"Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Alisa Liu, Noah A Smith, Daniel Khashabi, and Hannaneh Hajishirzi. 2022. Self-instruct: Aligning language models with self-generated instructions. arXiv preprint arXiv:2212.10560 (2022)."},{"key":"e_1_3_2_1_32_1","volume-title":"Less: Selecting influential data for targeted instruction tuning. arXiv preprint arXiv:2402.04333","author":"Xia Mengzhou","year":"2024","unstructured":"Mengzhou Xia, Sadhika Malladi, Suchin Gururangan, Sanjeev Arora, and Danqi Chen. 2024. Less: Selecting influential data for targeted instruction tuning. arXiv preprint arXiv:2402.04333 (2024)."},{"key":"e_1_3_2_1_33_1","volume-title":"Advances in Neural Information Processing Systems","volume":"36","author":"Yang Rui","year":"2024","unstructured":"Rui Yang, Lin Song, Yanwei Li, Sijie Zhao, Yixiao Ge, Xiu Li, and Ying Shan. 2024. Gpt4tools: Teaching large language model to use tools via self-instruction. Advances in Neural Information Processing Systems, Vol. 36 (2024)."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i16.17689"},{"key":"e_1_3_2_1_35_1","volume-title":"USNID: A framework for unsupervised and semi-supervised new intent discovery. CoRR, abs\/2304.07699","author":"Zhang Hanlei","year":"2023","unstructured":"Hanlei Zhang, Hua Xu, Xin Wang, Fei Long, and Kai Gao. 2023. USNID: A framework for unsupervised and semi-supervised new intent discovery. CoRR, abs\/2304.07699 (2023)."},{"key":"e_1_3_2_1_36_1","volume-title":"New intent discovery with pre-training and contrastive learning. arXiv preprint arXiv:2205.12914","author":"Zhang Yuwei","year":"2022","unstructured":"Yuwei Zhang, Haode Zhang, Li-Ming Zhan, Xiao-Ming Wu, and Albert Lam. 2022a. New intent discovery with pre-training and contrastive learning. arXiv preprint arXiv:2205.12914 (2022)."},{"key":"e_1_3_2_1_37_1","volume-title":"New intent discovery with pre-training and contrastive learning. arXiv preprint arXiv:2205.12914","author":"Zhang Yuwei","year":"2022","unstructured":"Yuwei Zhang, Haode Zhang, Li-Ming Zhan, Xiao-Ming Wu, and Albert Lam. 2022b. New intent discovery with pre-training and contrastive learning. arXiv preprint arXiv:2205.12914 (2022)."},{"key":"e_1_3_2_1_38_1","volume-title":"Advances in Neural Information Processing Systems","volume":"36","author":"Zhou Chunting","year":"2024","unstructured":"Chunting Zhou, Pengfei Liu, Puxin Xu, Srinivasan Iyer, Jiao Sun, Yuning Mao, Xuezhe Ma, Avia Efrat, Ping Yu, Lili Yu, et al., 2024. Lima: Less is more for alignment. Advances in Neural Information Processing Systems, Vol. 36 (2024)."}],"event":{"name":"WWW '26: The ACM Web Conference 2026","location":"Dubai United Arab Emirates","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the ACM Web Conference 2026"],"original-title":[],"deposited":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T16:36:21Z","timestamp":1775838981000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3774904.3792817"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,4,12]]},"references-count":38,"alternative-id":["10.1145\/3774904.3792817","10.1145\/3774904"],"URL":"https:\/\/doi.org\/10.1145\/3774904.3792817","relation":{},"subject":[],"published":{"date-parts":[[2026,4,12]]},"assertion":[{"value":"2026-04-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}