{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T18:34:05Z","timestamp":1742927645275,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":28,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819784868"},{"type":"electronic","value":"9789819784875"}],"license":[{"start":{"date-parts":[[2024,11,4]],"date-time":"2024-11-04T00:00:00Z","timestamp":1730678400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,4]],"date-time":"2024-11-04T00:00:00Z","timestamp":1730678400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-97-8487-5_23","type":"book-chapter","created":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T07:02:21Z","timestamp":1730617341000},"page":"323-337","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["AtomTool: Empowering Large Language Models with Tool Utilization Skills"],"prefix":"10.1007","author":[{"given":"Yongle","family":"Li","sequence":"first","affiliation":[]},{"given":"Zheng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Junqi","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Wenbo","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Yongyu","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Richang","family":"Hong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,4]]},"reference":[{"key":"23_CR1","unstructured":"Qin, Y., Hu, S., Lin, Y., et al.: Tool learning with foundation models (2023). arXiv:2304.08354"},{"key":"23_CR2","unstructured":"Hsieh, C.-Y., Chen, S.-A., Li, C.-L., Fujii, Y., Ratner, A., Lee, C.-Y., Krishna, R., Pfister, T.: Tool documentation enables zero-shot tool-usage with large language models (2023). arXiv:2308.00675"},{"key":"23_CR3","doi-asserted-by":"crossref","unstructured":"Liu, P., Yuan, W., Fu, J., Jiang, Z., Hayashi, H., Neubig, G.: Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 55(9), 1\u201335 (2023). ACM, New York, NY","DOI":"10.1145\/3560815"},{"key":"23_CR4","doi-asserted-by":"crossref","unstructured":"Zamfirescu-Pereira, J.D., Wong, R.Y., Hartmann, B., Yang, Q.: Why Johnny can\u2019t prompt: how non-AI experts try (and fail) to design LLM prompts. In: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pp. 1\u201321 (2023)","DOI":"10.1145\/3544548.3581388"},{"key":"23_CR5","doi-asserted-by":"crossref","unstructured":"Li, L., Zhang, Y., Chen, L.: Prompt distillation for efficient llm-based recommendation. In: Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, pp. 1348\u20131357 (2023)","DOI":"10.1145\/3583780.3615017"},{"key":"23_CR6","unstructured":"Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., et al.: Llama 2: Open foundation and fine-tuned chat models (2023). arXiv:2307.09288"},{"key":"23_CR7","unstructured":"Brown, T.B., Mann, B., Ryder, N., et al.: Language models are few-shot learners (2020). arXiv:2005.14165"},{"key":"23_CR8","unstructured":"Sun, T., Zhang, X., He, Z., Li, P., Cheng, Q., Yan, H., Liu, X., Shao, Y., Tang, Q., Zhao, X., et al.: Moss: Training conversational language models from synthetic data 7 (2023). arXiv:2307.15020"},{"key":"23_CR9","unstructured":"Hu, E.J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., Chen, W.: Lora: Low-rank adaptation of large language models (2021). arXiv:2106.09685"},{"key":"23_CR10","unstructured":"Taori, R., Gulrajani, I., Zhang, T., Dubois, Y., Li, X., Guestrin, C., Liang, P., Hashimoto, T.B.: Stanford Alpaca: An Instruction-following LLaMA Model (2023)"},{"key":"23_CR11","unstructured":"Chiang, W.-L., Li, Z., Lin, Z., Sheng, Y., Wu, Z., Zhang, H., Zheng, L., Zhuang, S., Zhuang, Y., Gonzalez, J.E., et al.: Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality. 2(3), 6 (2023). See https:\/\/vicuna.lmsys.org. Accessed 14 April 2023"},{"key":"23_CR12","first-page":"24824","volume":"35","author":"J Wei","year":"2022","unstructured":"Wei, J., Wang, X., Schuurmans, D., Bosma, M., Xia, F., Chi, E., Le, Q.V., Zhou, D., et al.: Chain-of-thought prompting elicits reasoning in large language models. Adv. Neural. Inf. Process. Syst. 35, 24824\u201324837 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"240","key":"23_CR13","first-page":"1","volume":"24","author":"A Chowdhery","year":"2023","unstructured":"Chowdhery, A., Narang, S., Devlin, J., Bosma, M., Mishra, G., Roberts, A., Barham, P., Chung, H.W., Sutton, C., Gehrmann, S., et al.: Palm: scaling language modeling with pathways. J. Mach. Learn. Res. 24(240), 1\u2013113 (2023)","journal-title":"J. Mach. Learn. Res."},{"key":"23_CR14","unstructured":"Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M.-A., Lacroix, T., Rozi\u00e8re, B., Goyal, N., Hambro, E., Azhar, F., et al.: Llama: Open and efficient foundation language models (2023). arXiv:2302.13971"},{"key":"23_CR15","unstructured":"Patil, S.G., Zhang, T., Wang, X., Gonzalez, J.E.: Gorilla: Large language model connected with massive apis (2023). arXiv:2305.15334"},{"key":"23_CR16","doi-asserted-by":"crossref","unstructured":"Komeili, M., Shuster, K., Weston, J.: Internet-augmented dialogue generation (2021). arXiv:2107.07566","DOI":"10.18653\/v1\/2022.acl-long.579"},{"key":"23_CR17","unstructured":"Lazaridou, A., Gribovskaya, E., Stokowiec, W., Grigorev, N.: Internet-augmented language models through few-shot prompting for open-domain question answering (2022). arXiv:2203.05115"},{"key":"23_CR18","unstructured":"Nakano, R., Hilton, J., Balaji, S., Wu, J., Ouyang, L., Kim, C., Hesse, C., Jain, S., Kosaraju, V., Saunders, W., et al.: Webgpt: Browser-assisted question-answering with human feedback (2021). arXiv:2112.09332"},{"key":"23_CR19","unstructured":"Gao, L., Madaan, A., Zhou, S., Alon, U., Liu, P., Yang, Y., Callan, J., Neubig, G.: Pal: Program-aided language models. In: International Conference on Machine Learning, pp. 10764\u201310799. PMLR (2023)"},{"key":"23_CR20","unstructured":"Borgeaud, S., Mensch, A., Hoffmann, J., Cai, T., Rutherford, E., Millican, K., Van Den Driessche, G.B., Lespiau, J.-B., Damoc, B., Clark, A., et al.: Improving language models by retrieving from trillions of tokens. In: International Conference on Machine Learning, pp. 2206\u20132240. PMLR (2022)"},{"key":"23_CR21","unstructured":"Guu, K., Lee, K., Tung, Z., Pasupat, P., Chang, M.: Retrieval augmented language model pre-training. In: International Conference on Machine Learning, pp. 3929\u20133938. PMLR (2020)"},{"key":"23_CR22","first-page":"9459","volume":"33","author":"P Lewis","year":"2020","unstructured":"Lewis, P., Perez, E., Piktus, A., Petroni, F., Karpukhin, V., Goyal, N., K\u00fcttler, H., Lewis, M., Yih, W.-T., Rockt\u00e4schel, T., et al.: Retrieval-augmented generation for knowledge-intensive nlp tasks. Adv. Neural. Inf. Process. Syst. 33, 9459\u20139474 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"23_CR23","first-page":"20744","volume":"35","author":"S Yao","year":"2022","unstructured":"Yao, S., Chen, H., Yang, J., Narasimhan, K.: Webshop: Towards scalable real-world web interaction with grounded language agents. Adv. Neural. Inf. Process. Syst. 35, 20744\u201320757 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"23_CR24","unstructured":"Paranjape, B., Lundberg, S., Singh, S., Hajishirzi, H., Zettlemoyer, L., Ribeiro, M.T.: Art: Automatic multi-step reasoning and tool-use for large language models (2023). arXiv:2303.09014"},{"key":"23_CR25","unstructured":"Cobbe, K., Kosaraju, V., Bavarian, M., Chen, M., Jun, H., Kaiser, L., Plappert, M., Tworek, J., Hilton, J., Nakano, R., et al.: Training verifiers to solve math word problems (2021). arXiv:2110.14168"},{"key":"23_CR26","unstructured":"Parisi, A., Zhao, Y., Fiedel, N.: Talm: Tool augmented language models (2022). arXiv:2205.12255"},{"key":"23_CR27","unstructured":"Chen, W., Ma, X., Wang, X., Cohen, W.W.: Program of thoughts prompting: Disentangling computation from reasoning for numerical reasoning tasks (2022). arXiv:2211.12588"},{"key":"23_CR28","unstructured":"Schick, T., Dwivedi-Yu, J., Dess\u00ec, R., Raileanu, R., Lomeli, M., Hambro, E., Zettlemoyer, L., Cancedda, N., Scialom, T.: Toolformer: Language models can teach themselves to use tools. Adv. Neural Inf. Process. Syst. 36 (2024)"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-8487-5_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T07:08:15Z","timestamp":1730617695000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-8487-5_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,4]]},"ISBN":["9789819784868","9789819784875"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-8487-5_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,4]]},"assertion":[{"value":"4 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Urumqi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2024.prcv.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}