{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T15:47:54Z","timestamp":1773330474287,"version":"3.50.1"},"reference-count":325,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T00:00:00Z","timestamp":1772841600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T00:00:00Z","timestamp":1772841600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Front. Comput. Sci."],"published-print":{"date-parts":[[2026,11]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>\n                    While large language models (LLMs) like ChatGPT have shown impressive capabilities in Natural Language Processing (NLP) tasks, a systematic investigation of their potential in this field remains largely unexplored. This study aims to address this gap by exploring the following questions. (1)\n                    <jats:italic>How are LLMs currently applied to NLP tasks in the literature<\/jats:italic>\n                    ? (2)\n                    <jats:italic>Have traditional NLP tasks already been solved with LLMs<\/jats:italic>\n                    ? (3)\n                    <jats:italic>What is the future of the LLMs for NLP<\/jats:italic>\n                    ? To answer these questions, we take the first step to provide a comprehensive overview of LLMs in NLP. Specifically, we first introduce a unified taxonomy including (1)\n                    <jats:italic>parameter-frozen paradigm<\/jats:italic>\n                    and (2)\n                    <jats:italic>parameter-tuning paradigm<\/jats:italic>\n                    to offer a unified perspective for understanding the current progress of LLMs in NLP. Furthermore, we summarize the new frontiers and the corresponding challenges, aiming to inspire further groundbreaking advancements. We hope this work offers valuable insights into {the potential and limitations} of LLMs, while also serving as a practical guide for building effective LLMs in NLP.\n                  <\/jats:p>","DOI":"10.1007\/s11704-025-50472-3","type":"journal-article","created":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T04:36:28Z","timestamp":1772858188000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Large language models meet NLP: a survey"],"prefix":"10.1007","volume":"20","author":[{"given":"Libo","family":"Qin","sequence":"first","affiliation":[]},{"given":"Qiguang","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Xiachong","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Yang","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Yongheng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Yinghui","family":"Li","sequence":"additional","affiliation":[]},{"given":"Min","family":"Li","sequence":"additional","affiliation":[]},{"given":"Wanxiang","family":"Che","sequence":"additional","affiliation":[]},{"given":"Philip S.","family":"Yu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,3,7]]},"reference":[{"key":"50472_CR1","unstructured":"Zhao WX, Zhou K, Li J, Tang T, Wang X, et al. A survey of large language models. 2023, arXiv preprint arXiv: 2303.18223"},{"key":"50472_CR2","unstructured":"Kaddour J, Harris J, Mozes M, Bradley H, Raileanu R, McHardy R. Challenges and applications of large language models. 2023, arXiv preprint arXiv: 2307.10169"},{"issue":"6","key":"50472_CR3","doi-asserted-by":"publisher","first-page":"160","DOI":"10.1145\/3649506","volume":"18","author":"J Yang","year":"2024","unstructured":"Yang J, Jin H, Tang R, Han X, Feng Q, Jiang H, Zhong S, Yin B, Hu X. Harnessing the power of LLMs in practice: a survey on chatgpt and beyond. ACM Transactions on Knowledge Discovery from Data, 2024, 18(6): 160","journal-title":"ACM Transactions on Knowledge Discovery from Data"},{"key":"50472_CR4","doi-asserted-by":"crossref","unstructured":"Hadi MU, Al Tashi Q, Qureshi R, Shah A, Muneer A, Irfan M, Zafar A, Shaikh MB, Akhtar N, Hassan SZ, Shoman M, Wu J, Mirjalili S, Shah M. Large language models: a comprehensive survey of its applications, challenges, limitations, and future prospects. 2023, TechRxiv","DOI":"10.36227\/techrxiv.23589741.v2"},{"key":"50472_CR5","first-page":"88","volume-title":"Proceedings of the 22nd Chinese National Conference on Computational Linguistics (Volume 2: Frontier Forum)","author":"Z Zhuang","year":"2023","unstructured":"Zhuang Z, Chen Q, Ma L, Li M, Han Y, Qian Y, Bai H, Zhang W, Liu T. Through the lens of core competency: survey on evaluation of large language models. In: Proceedings of the 22nd Chinese National Conference on Computational Linguistics (Volume 2: Frontier Forum). 2023, 88\u2013109"},{"key":"50472_CR6","unstructured":"Georgiev P, Lei VI, Burnell R, Bai L, Gulati A, et al. Gemini 1.5: unlocking multimodal understanding across millions of tokens of context. 2024, arXiv preprint arXiv: 2403.05530"},{"key":"50472_CR7","unstructured":"Guo D, Yang D, Zhang H, Song J, Zhang R, et al. DeepSeek-R1: incentivizing reasoning capability in LLMs via reinforcement learning. 2025, arXiv preprint arXiv: 2501.12948"},{"key":"50472_CR8","first-page":"24185","volume-title":"Proceedings of 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Z Chen","year":"2024","unstructured":"Chen Z, Wu J, Wang W, Su W, Chen G, Xing S, Zhong M, Zhang Q, Zhu X, Lu L, Li B, Luo P, Lu T, Qiao Y, Dai J. Intern VL: scaling up vision foundation models and aligning for generic visual-linguistic tasks. In: Proceedings of 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 2024, 24185\u201324198"},{"key":"50472_CR9","unstructured":"Chen Q, Yang M, Qin L, Liu J, Yan Z, Guan J, Peng D, Ji Y, Li H, Hu M, Zhang Y, Liang Y, Zhou Y, Wang J, Chen Z, Che W. AI4Research: a survey of artificial intelligence for scientific research. 2025, arXiv preprint arXiv: 2507.01903"},{"key":"50472_CR10","first-page":"159","volume-title":"Proceedings of the 34th International Conference on Neural Information Processing Systems","author":"TB Brown","year":"2020","unstructured":"Brown TB, Mann B, Ryder N, Subbiah M, Kaplan JD, et al. Language models are few-shot learners. In: Proceedings of the 34th International Conference on Neural Information Processing Systems. 2020, 159"},{"key":"50472_CR11","first-page":"2011","volume-title":"Proceedings of the 36th International Conference on Neural Information Processing Systems","author":"L Ouyang","year":"2022","unstructured":"Ouyang L, Wu J, Jiang X, Almeida D, Wainwright CL, Mishkin P, Zhang C, Agarwal S, Slama K, Ray A, Schulman J, Hilton J, Kelton F, Miller L, Simens M, Askell A, Welinder P, Christiano P, Leike J, Lowe R. Training language models to follow instructions with human feedback. In: Proceedings of the 36th International Conference on Neural Information Processing Systems. 2022, 2011"},{"issue":"1","key":"50472_CR12","first-page":"240","volume":"24","author":"A Chowdhery","year":"2023","unstructured":"Chowdhery A, Narang S, Devlin J, Bosma M, Mishra G, et al. PaLM: scaling language modeling with pathways. The Journal of Machine Learning Research, 2023, 24(1): 240","journal-title":"The Journal of Machine Learning Research"},{"key":"50472_CR13","unstructured":"Zhang S, Roller S, Goyal N, Artetxe M, Chen M, Chen S, Dewan C, Diab M, Li X, Lin XV, Mihaylov T, Ott M, Shleifer S, Shuster K, Simig D, Koura PS, Sridhar A, Wang T, Zettlemoyer L. OPT: open pre-trained transformer language models. 2022, arXiv preprint arXiv: 2205.01068"},{"key":"50472_CR14","unstructured":"Touvron H, Lavril T, Izacard G, Martinet X, Lachaux MA, Lacroix T, Rozi\u00e8re B, Goyal N, Hambro E, Azhar F, Rodriguez A, Joulin A, Grave E, Lample G. LLaMA: open and efficient foundation language models. 2023, arXiv preprint arXiv: 2302.13971"},{"key":"50472_CR15","volume-title":"Proceedings of the 10th International Conference on Learning Representations","author":"J Wei","year":"2022","unstructured":"Wei J, Bosma M, Zhao V, Guu K, Yu AW, Lester B, Du N, Dai AM, Le QV. Finetuned language models are zero-shot learners. In: Proceedings of the 10th International Conference on Learning Representations. 2022"},{"key":"50472_CR16","first-page":"1800","volume-title":"Proceedings of the 36th International Conference on Neural Information Processing Systems","author":"J Wei","year":"2022","unstructured":"Wei J, Wang X, Schuurmans D, Bosma M, Ichter B, Xia F, Chi EH, Le QV, Zhou D. Chain-of-thought prompting elicits reasoning in large language models. In: Proceedings of the 36th International Conference on Neural Information Processing Systems. 2022, 1800"},{"key":"50472_CR17","doi-asserted-by":"publisher","first-page":"11048","DOI":"10.18653\/v1\/2022.emnlp-main.759","volume-title":"Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing","author":"S Min","year":"2022","unstructured":"Min S, Lyu X, Holtzman A, Artetxe M, Lewis M, Hajishirzi H, Zettlemoyer L. Rethinking the role of demonstrations: what makes in-context learning work? In: Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing. 2022, 11048\u201311064"},{"key":"50472_CR18","unstructured":"Wei J, Tay Y, Bommasani R, Raffel C, Zoph B, Borgeaud S, Yogatama D, Bosma M, Zhou D, Metzler D, Chi EH, Hashimoto T, Vinyals O, Liang P, Dean J, Fedus W. Emergent abilities of large language models. Transactions on Machine Learning Research, 2022, 2022"},{"key":"50472_CR19","first-page":"1740","volume-title":"Proceedings of the 38th International Conference on Neural Information Processing Systems","author":"Q Chen","year":"2024","unstructured":"Chen Q, Qin L, Wang J, Zhou J, Che W. Unlocking the capabilities of thought: a reasoning boundary framework to quantify and optimize chain-of-thought. In: Proceedings of the 38th International Conference on Neural Information Processing Systems. 2024, 1740"},{"key":"50472_CR20","doi-asserted-by":"publisher","first-page":"12","DOI":"10.18653\/v1\/2023.newsum-1.2","volume-title":"Proceedings of the 4th New Frontiers in Summarization Workshop","author":"J Wang","year":"2023","unstructured":"Wang J, Liang Y, Meng F, Zou B, Li Z, Qu J, Zhou J. Zero-shot cross-lingual summarization via large language models. In: Proceedings of the 4th New Frontiers in Summarization Workshop. 2023, 12\u201323"},{"key":"50472_CR21","doi-asserted-by":"publisher","first-page":"8640","DOI":"10.18653\/v1\/2023.acl-long.482","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Y Wang","year":"2023","unstructured":"Wang Y, Zhang Z, Wang R. Element-aware summarization with large language models: expert-aligned evaluation and chain-of-thought method. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023, 8640\u20138665"},{"key":"50472_CR22","doi-asserted-by":"publisher","first-page":"16646","DOI":"10.18653\/v1\/2023.emnlp-main.1036","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"L Wang","year":"2023","unstructured":"Wang L, Lyu C, Ji T, Zhang Z, Yu D, Shi S, Tu Z. Document-level machine translation with large language models. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 16646\u201316661"},{"key":"50472_CR23","doi-asserted-by":"publisher","first-page":"5622","DOI":"10.18653\/v1\/2023.findings-emnlp.373","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023","author":"K Peng","year":"2023","unstructured":"Peng K, Ding L, Zhong Q, Shen L, Liu X, Zhang M, Ouyang Y, Tao D. Towards making the most of ChatGPT for machine translation. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023. 2023, 5622\u20135633"},{"key":"50472_CR24","unstructured":"Wei X, Cui X, Cheng N, Wang X, Zhang X, Huang S, Xie P, Xu J, Chen Y, Zhang M, Jiang Y, Han W. Zero-shot information extraction via chatting with ChatGPT. 2023, arXiv preprint arXiv: 2302.10205"},{"key":"50472_CR25","doi-asserted-by":"publisher","first-page":"3534","DOI":"10.18653\/v1\/2023.emnlp-main.214","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"Z Wan","year":"2023","unstructured":"Wan Z, Cheng F, Mao Z, Liu Q, Song H, Li J, Kurohashi S. GPT-RE: in-context learning for relation extraction using large language models. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 3534\u20133547"},{"key":"50472_CR26","unstructured":"Huang JT, Lam MH, Li EJ, Ren S, Wang W, Jiao W, Tu Z, Lyu MR. Emotionally numb or empathetic? Evaluating how LLMs feel using EmotionBench. 2023, arXiv preprint arXiv: 2308.03656"},{"key":"50472_CR27","unstructured":"Wang Z, Xie Q, Ding Z, Feng Y, Xia R. Is ChatGPT a good sentiment analyzer? A preliminary study. 2023, arXiv preprint arXiv: 2304.04339"},{"key":"50472_CR28","unstructured":"Chen Q, Qin L, Liu J, Peng D, Guan J, Wang P, Hu M, Zhou Y, Gao T, Che W. Towards reasoning era: a survey of long chain-of-thought for reasoning large language models. 2025, arXiv preprint arXiv: 2503.09567"},{"key":"50472_CR29","doi-asserted-by":"publisher","first-page":"9191","DOI":"10.18653\/v1\/2024.findings-acl.546","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: ACL 2024","author":"Y Zhang","year":"2024","unstructured":"Zhang Y, Chen Q, Li M, Che W, Qin L. AutoCAP: towards automatic cross-lingual alignment planning for zero-shot chain-of-thought. In: Proceedings of Findings of the Association for Computational Linguistics: ACL 2024. 2024, 9191\u20139200"},{"key":"50472_CR30","doi-asserted-by":"publisher","first-page":"103081","DOI":"10.1016\/j.inffus.2025.103081","volume":"120","author":"L Ren","year":"2025","unstructured":"Ren L, Liu Y, Ouyang C, Yu Y, Zhou S, He Y, Wan Y. DyLas: a dynamic label alignment strategy for large-scale multi-label text classification. Information Fusion, 2025, 120: 103081","journal-title":"Information Fusion"},{"key":"50472_CR31","first-page":"1613","volume-title":"Proceedings of the 36th International Conference on Neural Information Processing Systems","author":"T Kojima","year":"2022","unstructured":"Kojima T, Gu SS, Reid M, Matsuo Y, Iwasawa Y. Large language models are zero-shot reasoners. In: Proceedings of the 36th International Conference on Neural Information Processing Systems. 2022, 1613"},{"key":"50472_CR32","first-page":"2790","volume-title":"Proceedings of the 36th International Conference on Machine Learning","author":"N Houlsby","year":"2019","unstructured":"Houlsby N, Giurgiu A, Jastrzebski S, Morrone B, De Laroussilhe Q, Gesmundo A, Attariyan M, Gelly S. Parameter-efficient transfer learning for NLP. In: Proceedings of the 36th International Conference on Machine Learning. 2019, 2790\u20132799"},{"key":"50472_CR33","volume-title":"Proceedings of the 10th International Conference on Learning Representations","author":"EJ Hu","year":"2022","unstructured":"Hu EJ, Shen Y, Wallis P, Allen-Zhu Z, Li Y, Wang S, Wang L, Chen W. LoRA: low-rank adaptation of large language models. In: Proceedings of the 10th International Conference on Learning Representations. 2022"},{"key":"50472_CR34","first-page":"4582","volume-title":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)","author":"XL Li","year":"2021","unstructured":"Li XL, Liang P. Prefix-tuning: optimizing continuous prompts for generation. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 2021, 4582\u20134597"},{"key":"50472_CR35","first-page":"441","volume-title":"Proceedings of the 37th International Conference on Neural Information Processing Systems","author":"T Dettmers","year":"2023","unstructured":"Dettmers T, Pagnoni A, Holtzman A, Zettlemoyer L. QLORA: efficient finetuning of quantized LLMs. In: Proceedings of the 37th International Conference on Neural Information Processing Systems., 2023, 441"},{"key":"50472_CR36","doi-asserted-by":"publisher","first-page":"136","DOI":"10.1145\/3632410.3632463","volume-title":"Proceedings of the 7th Joint International Conference on Data Science & Management of Data (11th ACM IKDD CODS and 29th COMAD)","author":"N Mundra","year":"2024","unstructured":"Mundra N, Doddapaneni S, Dabre R, Kunchukuttan A, Puduppully R, Khapra MM. A comprehensive analysis of adapter efficiency. In: Proceedings of the 7th Joint International Conference on Data Science & Management of Data (11th ACM IKDD CODS and 29th COMAD). 2024, 136\u2013154"},{"issue":"7","key":"50472_CR37","doi-asserted-by":"publisher","first-page":"5731","DOI":"10.1007\/s10462-022-10144-1","volume":"55","author":"M Wankhade","year":"2022","unstructured":"Wankhade M, Chandra Sekhara Rao A, Kulkarni C. A survey on sentiment analysis methods, applications, and challenges. Artificial Intelligence Review, 2022, 55(7): 5731\u20135780","journal-title":"Artificial Intelligence Review"},{"key":"50472_CR38","first-page":"159","volume-title":"Proceedings of the 14th International Conference on Recent Advances in Natural Language Processing","author":"A Belkhir","year":"2023","unstructured":"Belkhir A, Sadat F. Beyond information: is ChatGPT empathetic enough? In Proceedings of the 14th International Conference on Recent Advances in Natural Language Processing. 2023, 159\u2013169"},{"key":"50472_CR39","doi-asserted-by":"publisher","first-page":"3881","DOI":"10.18653\/v1\/2024.findings-naacl.246","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: NAACL 2024","author":"W Zhang","year":"2024","unstructured":"Zhang W, Deng Y, Liu B, Pan SJ, Bing L. Sentiment analysis in the era of large language models: a reality check. In: Proceedings of Findings of the Association for Computational Linguistics: NAACL 2024. 2024, 3881\u20133906"},{"key":"50472_CR40","first-page":"298","volume-title":"Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"F Koto","year":"2024","unstructured":"Koto F, Beck T, Talat Z, Gurevych I, Baldwin T. Zero-shot sentiment analysis in low-resource languages using a multilingual sentiment lexicon. In: Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers). 2024, 298\u2013320"},{"key":"50472_CR41","first-page":"189","volume-title":"Proceedings of 2024 IEEE Conference on Artificial Intelligence (CAI)","author":"K Du","year":"2024","unstructured":"Du K, Xing F, Mao R, Cambria E. An evaluation of reasoning capabilities of large language models in financial sentiment analysis. In: Proceedings of 2024 IEEE Conference on Artificial Intelligence (CAI). 2024, 189\u2013194"},{"key":"50472_CR42","unstructured":"Zhao W, Zhao Y, Lu X, Wang S, Tong Y, Qin B. Is ChatGPT equipped with emotional dialogue capabilities? 2023, arXiv preprint arXiv: 2304.09582"},{"key":"50472_CR43","unstructured":"Xu X, Zhang JD, Xiao R, Xiong L. The limits of chatgpt in extracting aspect-category-opinion-sentiment quadruples: a comparative analysis. 2023, arXiv preprint arXiv: 2310.06502"},{"key":"50472_CR44","unstructured":"Lu Y, Ji Z, Du J, Shanqing Y, Xuan Q, Zhou T. From LLM-anation to LLM-orchestrator: coordinating small models for data labeling. 2025, arXiv preprint arXiv: 2506.16393"},{"key":"50472_CR45","unstructured":"Sun X, Li X, Zhang S, Wang S, Wu F, Li J, Zhang T, Wang G. Sentiment analysis through LLM negotiations. 2023, arXiv preprint arXiv: 2311.01876"},{"issue":"3","key":"50472_CR46","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1145\/3697009","volume":"34","author":"T Zhang","year":"2025","unstructured":"Zhang T, Irsan IC, Thung F, Lo D. Revisiting sentiment analysis for software engineering in the era of large language models. ACM Transactions on Software Engineering and Methodology, 2025, 34(3): 60","journal-title":"ACM Transactions on Software Engineering and Methodology"},{"key":"50472_CR47","doi-asserted-by":"publisher","first-page":"794","DOI":"10.18653\/v1\/2023.findings-acl.50","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: ACL 2023","author":"K Zhang","year":"2023","unstructured":"Zhang K, Gutierrez BJ, Su Y. Aligning instruction tasks unlocks large language models as zero-shot relation extractors. In: Proceedings of Findings of the Association for Computational Linguistics: ACL 2023. 2023, 794\u2013812"},{"key":"50472_CR48","doi-asserted-by":"publisher","first-page":"7935","DOI":"10.18653\/v1\/2023.emnlp-main.493","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"T Xie","year":"2023","unstructured":"Xie T, Li Q, Zhang J, Zhang Y, Liu Z, Wang H. Empirical study of zero-shot NER with ChatGPT. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 7935\u20137956, doi: https:\/\/doi.org\/10.18653\/v1\/2023.emnlp-main.493"},{"key":"50472_CR49","unstructured":"Li M, Zhang R. How far is language model from 100% few-shot named entity recognition in medical domain. 2023, arXiv preprint arXiv: 2307.00186"},{"key":"50472_CR50","doi-asserted-by":"publisher","first-page":"15339","DOI":"10.18653\/v1\/2023.acl-long.855","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"P Li","year":"2023","unstructured":"Li P, Sun T, Tang Q, Yan H, Wu Y, Huang X, Qiu X. CodeIE: large code generation models are better few-shot information extractors. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)., 2023, 15339\u201315353"},{"key":"50472_CR51","unstructured":"Bi Z, Chen J, Jiang Y, Xiong F, Guo W, Chen H, Zhang N. CodeKGC: code language model for generative knowledge graph construction. 2023, arXiv preprint arXiv: 2304.09048"},{"key":"50472_CR52","first-page":"456","volume-title":"Proceedings of the 7th International Conference on Natural Language and Speech Processing (ICNLSP 2024)","author":"R Fornasiere","year":"2024","unstructured":"Fornasiere R, Brunello N, Scotti V, Carman MJ. Medical information extraction with large language models. In: Proceedings of the 7th International Conference on Natural Language and Speech Processing (ICNLSP 2024). 2024, 456\u2013466"},{"key":"50472_CR53","doi-asserted-by":"crossref","unstructured":"Tang Y, Xiao Z, Li X, Fang Q, Zhang Q, Yee Tak Fong D, Tsz Tsun Lai F, Sze Ling Chui C, Wai Yin Chan E, Chi Kei Wong I. Large language model in medical information extraction from titles and abstracts with prompt engineering strategies: a comparative study of GPT-3.5 and GPT-4. 2024, MedRxiv","DOI":"10.1101\/2024.03.20.24304572"},{"key":"50472_CR54","unstructured":"Pan W, Chen Q, Xu X, Che W, Qin L. A preliminary evaluation of chatgpt for zero-shot dialogue understanding. 2023, arXiv preprint arXiv: 2304.04256"},{"key":"50472_CR55","first-page":"1109","volume-title":"Proceedings of the 24th Annual Conference of the International Speech Communication Association","author":"M He","year":"2023","unstructured":"He M, Garner PN. Can ChatGPT detect intent? Evaluating large language models for spoken language understanding. In: Proceedings of the 24th Annual Conference of the International Speech Communication Association., 2023, 1109\u20131113"},{"key":"50472_CR56","doi-asserted-by":"crossref","unstructured":"Hudecek V, Dusek O. Are LLMs all you need for task-oriented dialogue? 2023, arXiv preprint arXiv: 2304.06556","DOI":"10.18653\/v1\/2023.sigdial-1.21"},{"key":"50472_CR57","doi-asserted-by":"publisher","first-page":"936","DOI":"10.18653\/v1\/2023.acl-short.81","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)","author":"M Heck","year":"2023","unstructured":"Heck M, Lubis N, Ruppik B, Vukovic R, Feng S, Geishauser C, Lin HC, van Niekerk C, Gasic M. ChatGPT for zero-shot dialogue state tracking: a solution or an opportunity? In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 2023, 936\u2013950"},{"key":"50472_CR58","doi-asserted-by":"publisher","first-page":"14567","DOI":"10.63317\/4shdjhs9d9ud","volume-title":"Proceedings of 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)","author":"H Gao","year":"2024","unstructured":"Gao H, Lin TE, Li H, Yang M, Wu Y, Ma W, Huang F, Li Y. Self-explanation prompting improves dialogue understanding in large language models. In: Proceedings of 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024). 2024, 14567\u201314578"},{"key":"50472_CR59","doi-asserted-by":"publisher","first-page":"4330","DOI":"10.18653\/v1\/2022.findings-emnlp.318","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2022","author":"Z Li","year":"2022","unstructured":"Li Z, Chen W, Li S, Wang H, Qian J, Yan X. Controllable dialogue simulation with in-context learning. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2022. 2022, 4330\u20134347"},{"key":"50472_CR60","volume-title":"SEAGULL: an embodied agent for instruction following through situated dialog","author":"Y Zhang","year":"2023","unstructured":"Zhang Y, Yang J, Yu K, Dai Y, Storks S, Bao Y, Pan J, Devraj N, Ma Z, Chai J. SEAGULL: an embodied agent for instruction following through situated dialog. 2023"},{"key":"50472_CR61","doi-asserted-by":"publisher","first-page":"13348","DOI":"10.18653\/v1\/2023.findings-emnlp.891","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023","author":"X Zhang","year":"2023","unstructured":"Zhang X, Peng B, Li K, Zhou J, Meng H. SGP-TOD: building task bots effortlessly via schema-guided LLM prompting. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023. 2023, 13348\u201313369"},{"key":"50472_CR62","doi-asserted-by":"publisher","first-page":"11093","DOI":"10.18653\/v1\/2023.findings-emnlp.741","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023","author":"Y Wu","year":"2023","unstructured":"Wu Y, Dong G, Xu W. Semantic parsing by large language models for intricate updating strategies of zero-shot dialogue state tracking. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023. 2023, 11093\u201311099"},{"key":"50472_CR63","first-page":"14996","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: ACL 2024","author":"S Snigdha Sarathi Das","year":"2024","unstructured":"Snigdha Sarathi Das S, Shah C, Wan M, Neville J, Yang L, Andersen R, Buscher G, Safavi T. S3-DST: structured open-domain dialogue segmentation and state tracking in the era of LLMs. In: Proceedings of Findings of the Association for Computational Linguistics: ACL 2024. 2024, 14996\u201315014"},{"key":"50472_CR64","volume-title":"Dialogue distillery: crafting interpolable, interpretable, and introspectable dialogue from LLMs","author":"RA Chi","year":"2023","unstructured":"Chi RA, Kim J, Hickmann S, Li S, Chi G, Atchariyachanvanit T, Yu K, Chi NA, Dai G, Rammoorthy S, Wang JH, Sarthi P, Adams V, Xu BY, Xu BZ, Park K, Cao S, Manning CD. Dialogue distillery: crafting interpolable, interpretable, and introspectable dialogue from LLMs. 2023"},{"key":"50472_CR65","doi-asserted-by":"publisher","first-page":"2627","DOI":"10.18653\/v1\/2022.findings-emnlp.193","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2022","author":"Y Hu","year":"2022","unstructured":"Hu Y, Lee CH, Xie T, Yu T, Smith NA, Ostendorf M. In-context learning for few-shot dialogue state tracking. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2022. 2022, 2627\u20132643"},{"key":"50472_CR66","doi-asserted-by":"publisher","first-page":"5570","DOI":"10.18653\/v1\/2023.findings-acl.344","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: ACL 2023","author":"B King","year":"2023","unstructured":"King B, Flanigan J. Diverse retrieval-augmented in-context learning for dialogue state tracking. In: Proceedings of Findings of the Association for Computational Linguistics: ACL 2023. 2023, 5570\u20135585"},{"key":"50472_CR67","doi-asserted-by":"publisher","first-page":"229","DOI":"10.18653\/v1\/2023.sigdial-1.22","volume-title":"Proceedings of the 24th Annual Meeting of the Special Interest Group on Discourse and Dialogue","author":"A Addlesee","year":"2023","unstructured":"Addlesee A, Sieinska W, Gunson N, Garcia DH, Dondrup C, Lemon O. Multi-party goal tracking with LLMs: comparing pre-training, fine-tuning, and prompt engineering. In: Proceedings of the 24th Annual Meeting of the Special Interest Group on Discourse and Dialogue. 2023, 229\u2013241"},{"key":"50472_CR68","first-page":"1","volume-title":"Proceedings of the 2nd Workshop on Natural Language Interfaces","author":"W Chung","year":"2023","unstructured":"Chung W, Cahyawijaya S, Wilie B, Lovenia H, Fung P. InstructTODS: large language models for end-to-end task-oriented dialogue systems. In: Proceedings of the 2nd Workshop on Natural Language Interfaces. 2023, 1\u201321"},{"key":"50472_CR69","first-page":"1434","volume-title":"Proceedings of 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)","author":"CH Lee","year":"2024","unstructured":"Lee CH, Cheng H, Ostendorf M. OrchestraLLM: efficient orchestration of language models for dialogue state tracking. In: Proceedings of 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers). 2024, 1434\u20131445"},{"key":"50472_CR70","first-page":"545","volume-title":"Proceedings of the 24th International Symposium on Theory, Algorithmic Foundations, and Protocol Design for Mobile Networks and Mobile Computing","author":"E Lin","year":"2023","unstructured":"Lin E, Hale J, Gratch J. Toward a better understanding of the emotional dynamics of negotiation with large language models. In: Proceedings of the 24th International Symposium on Theory, Algorithmic Foundations, and Protocol Design for Mobile Networks and Mobile Computing. 2023, 545\u2013550"},{"key":"50472_CR71","unstructured":"Cao L. DiagGPT: an LLM-based chatbot with automatic topic management for task-oriented dialogue. 2023, arXiv preprint arXiv: 2308.08043"},{"key":"50472_CR72","volume-title":"Proceedings of NeurIPS 2023 Second Table Representation Learning Workshop","author":"A Singha","year":"2023","unstructured":"Singha A, Cambronero J, Gulwani S, Le V, Parnin C. Tabular representation, noisy operators, and impacts on table structure understanding tasks in LLMs. In: Proceedings of NeurIPS 2023 Second Table Representation Learning Workshop. 2023"},{"key":"50472_CR73","volume-title":"Proceedings of the 12th International Conference on Learning Representations","author":"S Patnaik","year":"2024","unstructured":"Patnaik S, Changwal H, Aggarwal M, Bhatia S, Kumar Y, Krishnamurthy B. CABINET: content relevance-based noise reduction for table question answering. In: Proceedings of the 12th International Conference on Learning Representations. 2024"},{"key":"50472_CR74","doi-asserted-by":"publisher","first-page":"174","DOI":"10.1145\/3539618.3591708","volume-title":"Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval","author":"Y Ye","year":"2023","unstructured":"Ye Y, Hui B, Yang M, Li B, Huang F, Li Y. Large language models are versatile decomposers: decomposing evidence and questions for table-based reasoning. In: Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2023, 174\u2013184"},{"key":"50472_CR75","first-page":"575","volume-title":"Proceedings of the 16th Asian Conference on Machine Learning","author":"J Ye","year":"2025","unstructured":"Ye J, Du M, Wang G. Dataframe QA: a universal LLM framework on dataframe question answering without data exposure. In: Proceedings of the 16th Asian Conference on Machine Learning. 2025, 575\u2013590"},{"key":"50472_CR76","unstructured":"Sui Y, Zhou M, Zhou M, Han S, Zhang D. GPT4Table: can large language models understand structured table data? A benchmark and empirical study. 2023, arXiv preprint arXiv: 2305.13062"},{"key":"50472_CR77","doi-asserted-by":"publisher","first-page":"10306","DOI":"10.18653\/v1\/2024.findings-emnlp.603","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2024","author":"Y Sui","year":"2024","unstructured":"Sui Y, Zou J, Zhou M, He X, Du L, Han S, Zhang D. TAP4LLM: table provider on sampling, augmenting, and packing semi-structured data for large language model reasoning. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2024. 2024, 10306\u201310323"},{"key":"50472_CR78","volume-title":"Proceedings of the 11th International Conference on Learning Representations","author":"Z Cheng","year":"2023","unstructured":"Cheng Z, Xie T, Shi P, Li C, Nadkarni R, Hu Y, Xiong C, Radev D, Ostendorf M, Zettlemoyer L, Smith NA, Yu T. Binding language models in symbolic languages. In: Proceedings of the 11th International Conference on Learning Representations. 2023"},{"key":"50472_CR79","unstructured":"Zhang W, Shen Y, Lu W, Zhuang Y. Data-copilot: bridging billions of data and humans with autonomous workflow. 2023, arXiv preprint arXiv: 2306.07209"},{"key":"50472_CR80","doi-asserted-by":"publisher","first-page":"2131","DOI":"10.18653\/v1\/2023.emnlp-main.132","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"Z Zhang","year":"2023","unstructured":"Zhang Z, Li X, Gao Y, Lou JG. CRT-QA: a dataset of complex reasoning question answering over tabular data. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 2131\u20132153, doi: https:\/\/doi.org\/10.18653\/v1\/2023.emnlp-main.132"},{"issue":"8","key":"50472_CR81","doi-asserted-by":"publisher","first-page":"1981","DOI":"10.14778\/3659437.3659452","volume":"17","author":"Y Zhang","year":"2024","unstructured":"Zhang Y, Henkel J, Floratou A, Cahoon J, Deep S, Patel JM. ReAcTable: enhancing react for table question answering. Proceedings of the VLDB Endowment, 2024, 17(8): 1981\u20131994","journal-title":"Proceedings of the VLDB Endowment"},{"key":"50472_CR82","first-page":"3086","volume-title":"Proceedings of the 27th International Conference on Computer Supported Cooperative Work in Design","author":"H Zhang","year":"2024","unstructured":"Zhang H, Si Q, Fu P, Lin Z, Wang W. Are large language models table-based fact-checkers? In: Proceedings of the 27th International Conference on Computer Supported Cooperative Work in Design. 2024, 3086\u20133091"},{"key":"50472_CR83","doi-asserted-by":"publisher","first-page":"1120","DOI":"10.18653\/v1\/2023.findings-eacl.83","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EACL 2023","author":"W Chen","year":"2023","unstructured":"Chen W. Large language models are few(1)-shot table reasoners. In: Proceedings of Findings of the Association for Computational Linguistics: EACL 2023. 2023, 1120\u20131130"},{"key":"50472_CR84","unstructured":"Luo T, Lei F, Lei J, Liu W, He S, Zhao J, Liu K. HRoT: hybrid prompt strategy and retrieval of thought for table-text hybrid question answering. 2023, arXiv preprint arXiv: 2309.12669"},{"key":"50472_CR85","first-page":"220","volume-title":"Proceedings of the 37th International Conference on Neural Information Processing Systems","author":"H Li","year":"2023","unstructured":"Li H, Su J, Chen Y, Li Q, Zhang Z. SheetCopilot: bringing software productivity to the next level through large language models. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. 2023, 220"},{"key":"50472_CR86","doi-asserted-by":"publisher","first-page":"9237","DOI":"10.18653\/v1\/2023.emnlp-main.574","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"J Jiang","year":"2023","unstructured":"Jiang J, Zhou K, Dong Z, Ye K, Zhao X, Wen JR. StructGPT: a general framework for large language model to reason over structured data. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 9237\u20139251, doi: https:\/\/doi.org\/10.18653\/v1\/2023.emnlp-main.574"},{"key":"50472_CR87","volume-title":"Proceedings of the 12th International Conference on Learning Representations","author":"Z Wang","year":"2024","unstructured":"Wang Z, Zhang H, Li CL, Eisenschlos JM, Perot V, Wang Z, Miculicich L, Fujii Y, Shang J, Lee CY, Pfister T. Chain-of-table: evolving tables in the reasoning chain for table understanding. In: Proceedings of the 12th International Conference on Learning Representations. 2024"},{"key":"50472_CR88","volume-title":"Proceedings of the 12th International Conference on Learning Representations","author":"K Kong","year":"2024","unstructured":"Kong K, Zhang J, Shen Z, Srinivasan B, Lei C, Faloutsos C, Rangwala H, Karypis G. OpenTab: advancing large language models as open-domain table reasoners. In: Proceedings of the 12th International Conference on Learning Representations. 2024"},{"key":"50472_CR89","unstructured":"Goyal T, Li JJ, Durrett G. News summarization and evaluation in the era of GPT-3. 2022, arXiv preprint arXiv: 2209.12356"},{"key":"50472_CR90","first-page":"2764","volume-title":"Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"M Ravaut","year":"2023","unstructured":"Ravaut M, Sun A, Chen NF, Joty S. On context utilization in summarization with large language models. In: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023, 2764\u20132781"},{"key":"50472_CR91","doi-asserted-by":"publisher","first-page":"9282","DOI":"10.18653\/v1\/2023.findings-acl.591","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: ACL 2023","author":"A Bhaskar","year":"2023","unstructured":"Bhaskar A, Fabbri AR, Durrett G. Prompted opinion summarization with GPT-3.5. In: Proceedings of Findings of the Association for Computational Linguistics: ACL 2023. 2023, 9282\u20139300"},{"key":"50472_CR92","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1162\/tacl_a_00632","volume":"12","author":"T Zhang","year":"2023","unstructured":"Zhang T, Ladhak F, Durmus E, Liang P, McKeown K, Hashimoto TB. Benchmarking large language models for news summarization. Transactions of the Association for Computational Linguistics, 2023, 12: 39\u201357","journal-title":"Transactions of the Association for Computational Linguistics"},{"key":"50472_CR93","doi-asserted-by":"publisher","first-page":"3270","DOI":"10.18653\/v1\/2023.findings-emnlp.214","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023","author":"H Zhang","year":"2023","unstructured":"Zhang H, Liu X, Zhang J. Extractive summarization via chatgpt for faithful summary generation. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023. 2023, 3270\u20133278"},{"key":"50472_CR94","doi-asserted-by":"crossref","unstructured":"Adams G, Fabbri A, Ladhak F, Lehman E, Elhadad N. From sparse to dense: GPT-4 summarization with chain of density prompting. 2023, arXiv preprint arXiv: 2309.04269","DOI":"10.18653\/v1\/2023.newsum-1.7"},{"key":"50472_CR95","doi-asserted-by":"publisher","first-page":"56","DOI":"10.18653\/v1\/2023.newsum-1.6","volume-title":"Proceedings of the 4th New Frontiers in Summarization Workshop","author":"Y Tang","year":"2023","unstructured":"Tang Y, Puduppully R, Liu Z, Chen N. In-context learning of large language models for controlled dialogue summarization: a holistic benchmark and empirical analysis. In: Proceedings of the 4th New Frontiers in Summarization Workshop. 2023, 56\u201367, doi: https:\/\/doi.org\/10.18653\/v1\/2023.newsum-1.6"},{"key":"50472_CR96","unstructured":"Chen M, Tworek J, Jun H, Yuan Q, Ponde de Oliveira Pinto H, et al. Evaluating large language models trained on code. 2021, arXiv preprint arXiv: 2107.03374"},{"key":"50472_CR97","volume-title":"Proceedings of the 11th International Conference on Learning Representations","author":"E Nijkamp","year":"2023","unstructured":"Nijkamp E, Pang B, Hayashi H, Tu L, Wang H, Zhou Y, Savarese S, Xiong C. CodeGen: an open large language model for code with multi-turn program synthesis. In: Proceedings of the 11th International Conference on Learning Representations. 2023"},{"key":"50472_CR98","unstructured":"Christopoulou F, Lampouras G, Gritta M, Zhang G, Guo Y, et al. PanGu-coder: program synthesis with function-level language modeling. 2022, arXiv preprint arXiv: 2207.11280"},{"key":"50472_CR99","volume-title":"Proceedings of the 12th International Conference on Learning Representations","author":"Z Luo","year":"2024","unstructured":"Luo Z, Xu C, Zhao P, Sun Q, Geng X, Hu W, Tao C, Ma J, Lin Q, Jiang D. WizardCoder: Empowering code large language models with evol-instruct. In: Proceedings of the 12th International Conference on Learning Representations. 2024"},{"key":"50472_CR100","unstructured":"Allal LB, Li R, Kocetkov D, Mou C, Akiki C, et al. SantaCoder: don\u2019t reach for the stars! 2023, arXiv preprint arXiv: 2301.03988"},{"key":"50472_CR101","unstructured":"Li R, Ben Allal L, Zi Y, Muennighoff N, Kocetkov D, et al. StarCoder: may the source be with you! Transactions on Machine Learning Research, 2023, 2023"},{"key":"50472_CR102","unstructured":"Li Y, Bubeck S, Eldan R, Del Giorno A, Gunasekar S, Lee YT. Textbooks are all you need II: phi-1.5 technical report. 2023, arXiv preprint arXiv: 2309.05463"},{"key":"50472_CR103","unstructured":"Guo D, Zhu Q, Yang D, Xie Z, Dong K, Zhang W, Chen G, Bi X, Wu Y, Li YK, Luo F, Xiong Y, Liang W. DeepSeek-coder: when the large language model meets programming \u2014 the rise of code intelligence. 2024, arXiv preprint arXiv: 2401.14196"},{"key":"50472_CR104","unstructured":"Roziere B, Gehring J, Gloeckle F, Sootla S, Gat I, et al. Code llama: open foundation models for code. 2023, arXiv preprint arXiv: 2308.12950"},{"key":"50472_CR105","doi-asserted-by":"crossref","unstructured":"Zheng Q, Xia X, Zou X, Dong Y, Wang S, Xue Y, Wang Z, Shen L, Wang A, Li Y, Su T, Yang Z, Tang J. CodeGeeX: a pre-trained model for code generation with multilingual evaluations on HumanEval-X. 2023, arXiv preprint arXiv: 2303.17568","DOI":"10.1145\/3580305.3599790"},{"key":"50472_CR106","unstructured":"Wei X, Wei H, Lin H, Li T, Zhang P, Ren X, Li M, Wan Y, Cao Z, Xie B, Hu T, Li S, Hui B, Yu B, Liu D, Yang B, Huang F, Xie J. PolyLM: an open source polyglot large language model. 2023, arXiv preprint arXiv: 2307.06018"},{"key":"50472_CR107","unstructured":"Zhu W, Lv Y, Dong Q, Yuan F, Xu J, Huang S, Kong L, Chen J, Li L. Extrapolating large language models to non-english by aligning languages. 2023, arXiv preprint arXiv: 2308.04948"},{"key":"50472_CR108","doi-asserted-by":"publisher","first-page":"8616","DOI":"10.18653\/v1\/2023.emnlp-main.532","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"C Li","year":"2023","unstructured":"Li C, Liu M, Zhang H, Chen Y, Xu J, Zhou M. MT2: towards a multi-task machine translation model with translation-specific in-context learning. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 8616\u20138627"},{"key":"50472_CR109","unstructured":"Li J, Tang Z, Ding Y, Wang P, Guo P, You W, Qiao D, Chen W, Fu G, Zhu Q, Zhou G, Zhang M. OpenBA: an open-sourced 15B bilingual asymmetric seq2seq model pre-trained from scratch. 2023, arXiv preprint arXiv: 2309.10706"},{"key":"50472_CR110","doi-asserted-by":"publisher","first-page":"11127","DOI":"10.18653\/v1\/2023.findings-emnlp.744","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023","author":"DM Alves","year":"2023","unstructured":"Alves DM, Guerreiro NM, Alves J, Pombal J, Rei R, de Souza J, Colombo P, Martins A. Steering large language models for machine translation with finetuning and in-context learning. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023. 2023, 11127\u201311148"},{"key":"50472_CR111","unstructured":"Raunak V, Awadalla HH, Menezes A. Dissecting in-context learning of translations in GPTs. 2023, arXiv preprint arXiv: 2310.15987"},{"key":"50472_CR112","doi-asserted-by":"publisher","first-page":"958","DOI":"10.18653\/v1\/2024.emnlp-main.55","volume-title":"Proceedings of 2024 Conference on Empirical Methods in Natural Language Processing","author":"H Lu","year":"2024","unstructured":"Lu H, Yang H, Huang H, Zhang D, Lam W, Wei F. Chain-of-dictionary prompting elicits translation in large language models. In: Proceedings of 2024 Conference on Empirical Methods in Natural Language Processing. 2024, 958\u2013976"},{"key":"50472_CR113","volume-title":"Proceedings of the 11th International Conference on Learning Representations","author":"Z Zhang","year":"2023","unstructured":"Zhang Z, Zhang A, Li M, Smola A. Automatic chain of thought prompting in large language models. In: Proceedings of the 11th International Conference on Learning Representations. 2023"},{"key":"50472_CR114","volume-title":"Proceedings of the 11th International Conference on Learning Representations","author":"X Wang","year":"2023","unstructured":"Wang X, Wei J, Schuurmans D, Le QV, Chi EH, Narang S, Chowdhery A, Zhou D. Self-consistency improves chain of thought reasoning in language models. In: Proceedings of the 11th International Conference on Learning Representations. 2023"},{"key":"50472_CR115","volume-title":"Proceedings of the 11th International Conference on Learning Representations","author":"P Lu","year":"2023","unstructured":"Lu P, Qiu L, Chang KW, Wu YN, Zhu SC, Rajpurohit T, Clark P, Kalyan A. Dynamic prompt learning via policy gradient for semi-structured mathematical reasoning. In: Proceedings of the 11th International Conference on Learning Representations. 2023"},{"key":"50472_CR116","first-page":"10764","volume-title":"Proceedings of the 40th International Conference on Machine Learning","author":"L Gao","year":"2023","unstructured":"Gao L, Madaan A, Zhou S, Alon U, Liu P, Yang Y, Callan J, Neubig G. PAL: program-aided language models. In: Proceedings of the 40th International Conference on Machine Learning. 2023, 10764\u201310799"},{"key":"50472_CR117","first-page":"942","volume-title":"Proceedings of 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)","author":"D Das","year":"2024","unstructured":"Das D, Banerjee D, Aditya S, Kulkarni A. MATHSENSEI: a tool-augmented large language model for mathematical reasoning. In: Proceedings of 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers). 2024, 942\u2013966"},{"key":"50472_CR118","unstructured":"Wang Z, Xia R, Yu J. UnifiedABSA: a unified ABSA framework based on multi-task instruction tuning. 2022, arXiv preprint arXiv: 2211.10986"},{"key":"50472_CR119","doi-asserted-by":"publisher","first-page":"19","DOI":"10.18653\/v1\/2023.wassa-1.3","volume-title":"Proceedings of the 13th Workshop on Computational Approaches to Subjectivity, Sentiment, & Social Media Analysis","author":"S Varia","year":"2023","unstructured":"Varia S, Wang S, Halder K, Vacareanu R, Ballesteros M, Benajiba Y, John NA, Anubhai R, Muresan S, Roth D. Instruction tuning for few-shot aspect-based sentiment analysis. In: Proceedings of the 13th Workshop on Computational Approaches to Subjectivity, Sentiment, & Social Media Analysis. 2023, 19\u201327"},{"key":"50472_CR120","doi-asserted-by":"publisher","first-page":"6062","DOI":"10.18653\/v1\/2023.findings-emnlp.403","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023","author":"B Yang","year":"2023","unstructured":"Yang B, Li J. Visual elements mining as prompts for instruction learning for target-oriented multimodal sentiment classification. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023. 2023, 6062\u20136075"},{"issue":"3","key":"50472_CR121","doi-asserted-by":"publisher","first-page":"595","DOI":"10.1162\/COLI_r_00259","volume":"42","author":"J Zhao","year":"2016","unstructured":"Zhao J, Liu K, Xu L. Sentiment analysis: mining opinions, sentiments, and emotions. Computational Linguistics, 2016, 42(3): 595\u2013598","journal-title":"Computational Linguistics"},{"key":"50472_CR122","doi-asserted-by":"publisher","first-page":"615","DOI":"10.18653\/v1\/2024.findings-emnlp.34","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2024","author":"H Qiu","year":"2024","unstructured":"Qiu H, He H, Zhang S, Li A, Lan Z. SMILE: single-turn to multi-turn inclusive language expansion via ChatGPT for mental health support. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2024. 2024, 615\u2013636"},{"key":"50472_CR123","doi-asserted-by":"publisher","first-page":"1666","DOI":"10.18653\/v1\/2023.acl-short.143","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)","author":"D Lu","year":"2023","unstructured":"Lu D, Ran S, Tetreault J, Jaimes A. Event extraction as question generation and answering. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 2023, 1666\u20131688"},{"key":"50472_CR124","unstructured":"Gan C, Zhang Q, Mori T. GIELLM: Japanese general information extraction large language model utilizing mutual reinforcement effect. 2023, arXiv preprint arXiv: 2311.06838"},{"key":"50472_CR125","volume-title":"Proceedings of the 12th International Conference on Learning Representations","author":"O Sainz","year":"2024","unstructured":"Sainz O, Garc\u00eda-Ferrero I, Agerri R, Lopez de Lacalle O, Rigau G, Agirre E. GoLLIE: annotation guidelines improve zero-shot information-extraction. In: Proceedings of the 12th International Conference on Learning Representations. 2024"},{"key":"50472_CR126","unstructured":"Wang X, Zhou W, Zu C, Xia H, Chen T, Zhang Y, Zheng R, Ye J, Zhang Q, Gui T, Kang J, Yang J, Li S, Du C. InstructUIE: multi-task instruction tuning for unified information extraction. 2023, arXiv preprint arXiv: 2304.08085"},{"key":"50472_CR127","doi-asserted-by":"publisher","first-page":"6998","DOI":"10.18653\/v1\/2023.emnlp-main.433","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"S Snigdha Sarathi Das","year":"2023","unstructured":"Snigdha Sarathi Das S, Zhang RH, Shi P, Yin W, Zhang R. Unified low-resource sequence labeling by sample-aware dynamic sparse finetuning. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 6998\u20137010, doi: https:\/\/doi.org\/10.18653\/v1\/2023.emnlp-main.433"},{"key":"50472_CR128","doi-asserted-by":"publisher","first-page":"277","DOI":"10.18653\/v1\/2023.acl-long.16","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Z Liang","year":"2023","unstructured":"Liang Z, Wei F, Jie Y, Qian Y, Hao Z, Han B. Prompts can play lottery tickets well: achieving lifelong information extraction via lottery prompt tuning. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023, 277\u2013292, doi: https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.16"},{"issue":"1","key":"50472_CR129","doi-asserted-by":"publisher","first-page":"1418","DOI":"10.1038\/s41467-024-45563-x","volume":"15","author":"J Dagdelen","year":"2024","unstructured":"Dagdelen J, Dunn A, Lee S, Walker N, Rosen AS, Ceder G, Persson KA, Jain A. Structured information extraction from scientific text with large language models. Nature Communications, 2024, 15(1): 1418","journal-title":"Nature Communications"},{"key":"50472_CR130","doi-asserted-by":"publisher","first-page":"211","DOI":"10.18653\/v1\/2024.acl-demos.20","volume-title":"Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 3: System Demonstrations)","author":"L Xue","year":"2024","unstructured":"Xue L, Zhang D, Dong Y, Tang J. AutoRE: document-level relation extraction with large language models. In: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 3: System Demonstrations). 2024, 211\u2013220"},{"key":"50472_CR131","doi-asserted-by":"publisher","DOI":"10.31219\/osf.io\/82hac","volume-title":"Interleaved multi-modal document representations for large-scale information retrieval using large language models","author":"D Rixewa","year":"2024","unstructured":"Rixewa D, Anderson K, Dubois L, Harrington M. Interleaved multi-modal document representations for large-scale information retrieval using large language models. 2024"},{"key":"50472_CR132","doi-asserted-by":"publisher","first-page":"602","DOI":"10.18653\/v1\/2022.emnlp-main.39","volume-title":"Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing","author":"T Xie","year":"2022","unstructured":"Xie T, Wu CH, Shi P, Zhong R, Scholak T, et al. UnifiedSKG: unifying and multi-tasking structured knowledge grounding with text-to-text language models. In: Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing. 2022, 602\u2013631"},{"key":"50472_CR133","unstructured":"Zhao J, Gupta R, Cao Y, Yu D, Wang M, Lee H, Rastogi A, Shafran I, Wu Y. Description-driven task-oriented dialog modeling. 2022, arXiv preprint arXiv: 2201.08904"},{"key":"50472_CR134","first-page":"4541","volume-title":"Proceedings of 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","author":"R Gupta","year":"2022","unstructured":"Gupta R, Lee H, Zhao J, Cao Y, Rastogi A, Wu Y. Show, don\u2019t tell: demonstrations outperform descriptions for schema-guided task-oriented dialogue. In: Proceedings of 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2022, 4541\u20134549"},{"key":"50472_CR135","doi-asserted-by":"publisher","first-page":"3428","DOI":"10.18653\/v1\/2022.findings-emnlp.250","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2022","author":"D Yu","year":"2022","unstructured":"Yu D, Wang M, Cao Y, El Shafey L, Shafran I, Soltau H. Knowledge-grounded dialog state tracking. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2022. 2022, 3428\u20133435"},{"key":"50472_CR136","doi-asserted-by":"publisher","first-page":"739","DOI":"10.18653\/v1\/2023.emnlp-main.48","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"Y Feng","year":"2023","unstructured":"Feng Y, Lu Z, Liu B, Zhan L, Wu XM. Towards LLM-driven dialogue state tracking. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 739\u2013755"},{"key":"50472_CR137","first-page":"1","volume-title":"Proceedings of 2023 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU)","author":"H Liu","year":"2023","unstructured":"Liu H, Cai Y, Zhou Y, Ou Z, Huang Y, Feng J. Prompt pool based class-incremental continual learning for dialog state tracking. In: Proceedings of 2023 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU). 2023, 1\u20138"},{"key":"50472_CR138","unstructured":"Li P, He Y, Yashar D, Cui W, Ge S, Zhang H, Fainman DR, Zhang D, Chaudhuri S. Table-GPT: table-tuned GPT for diverse table tasks. 2023, arXiv preprint arXiv: 2310.09263"},{"key":"50472_CR139","unstructured":"Xue S, Jiang C, Shi W, Cheng F, Chen K, Yang H, Zhang Z, He J, Zhang H, Wei G, Zhao W, Zhou F, Qi D, Yi H, Liu S, Chen F. DB-GPT: empowering database interactions with private large language models. 2023, arXiv preprint arXiv: 2312.17449"},{"key":"50472_CR140","unstructured":"Zhang H, Dong Y, Xiao C, Oyamada M. Jellyfish: a large language model for data preprocessing. 2023, arXiv preprint arXiv: 2312.01678"},{"key":"50472_CR141","doi-asserted-by":"crossref","unstructured":"Zhu F, Liu Z, Feng F, Wang C, Li M, Chua TS. TAT-LLM: a specialized language model for discrete reasoning over tabular and textual data. 2024, arXiv preprint arXiv: 2401.13223","DOI":"10.1145\/3677052.3698685"},{"key":"50472_CR142","doi-asserted-by":"publisher","first-page":"10252","DOI":"10.18653\/v1\/2024.findings-emnlp.600","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2024","author":"F Bai","year":"2024","unstructured":"Bai F, Kang J, Stanovsky G, Freitag D, Ritter A. Schema-driven information extraction from heterogeneous tables. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2024, 2024, 10252\u201310273"},{"key":"50472_CR143","first-page":"6024","volume-title":"Proceedings of 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)","author":"T Zhang","year":"2024","unstructured":"Zhang T, Yue X, Li Y, Sun H. TableLlama: towards open large generalist models for tables. In: Proceedings of 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers). 2024, 6024\u20136044"},{"key":"50472_CR144","doi-asserted-by":"publisher","first-page":"22376","DOI":"10.18653\/v1\/2025.acl-long.1090","volume-title":"Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"X He","year":"2025","unstructured":"He X, Liu Y, Zhou M, He Y, Dong H, Han S, Yuan Z, Zhang D. TableLoRA: low-rank adaptation on table structure understanding for large language models. In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2025, 22376\u201322391"},{"issue":"3","key":"50472_CR145","doi-asserted-by":"publisher","first-page":"176","DOI":"10.1145\/3654979","volume":"2","author":"P Li","year":"2024","unstructured":"Li P, He Y, Yashar D, Cui W, Ge S, Zhang H, Fainman DR, Zhang D, Chaudhuri S. Table-GPT: table fine-tuned GPT for diverse table tasks. Proceedings of the ACM on Management of Data, 2024, 2(3): 176","journal-title":"Proceedings of the ACM on Management of Data"},{"key":"50472_CR146","doi-asserted-by":"publisher","first-page":"12737","DOI":"10.18653\/v1\/2023.acl-long.713","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"A Pagnoni","year":"2023","unstructured":"Pagnoni A, Fabbri AR, Kryscinski W, Wu CS. Socratic pretraining: question-driven pretraining for controllable summarization. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023, 12737\u201312755"},{"key":"50472_CR147","doi-asserted-by":"publisher","first-page":"4848","DOI":"10.18653\/v1\/2022.naacl-main.357","volume-title":"Proceedings of 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","author":"L Zhao","year":"2022","unstructured":"Zhao L, Zheng F, Zeng W, He K, Xu W, Jiang H, Wu W, Wu Y. Domain-oriented prefix-tuning: towards efficient and generalizable fine-tuning for zero-shot dialogue summarization. In: Proceedings of 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2022, 4848\u20134862, doi: https:\/\/doi.org\/10.18653\/v1\/2022.naacl-main.357"},{"key":"50472_CR148","doi-asserted-by":"publisher","first-page":"3704","DOI":"10.18653\/v1\/2022.emnlp-main.243","volume-title":"Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing","author":"R Yuan","year":"2022","unstructured":"Yuan R, Wang Z, Cao Z, Li W. Few-shot query-focused summarization with prefix-merging. In: Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing. 2022, 3704\u20133714"},{"key":"50472_CR149","doi-asserted-by":"publisher","first-page":"3694","DOI":"10.1109\/TASLP.2024.3414313","volume":"32","author":"X Feng","year":"2024","unstructured":"Feng X, Feng X, Du X, Kan MY, Qin B. Adapter-based selective knowledge distillation for federated multi-domain meeting summarization. IEEE\/ACM Transactions on Audio, Speech, and Language Processing, 2024, 32: 3694\u20133708","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"50472_CR150","unstructured":"Ravaut M, Chen H, Zhao R, Qin C, Joty S, Chen N. PromptSum: parameter-efficient controllable abstractive summarization. 2023, arXiv preprint arXiv: 2308.03117"},{"key":"50472_CR151","doi-asserted-by":"publisher","first-page":"8696","DOI":"10.18653\/v1\/2021.emnlp-main.685","volume-title":"Proceedings of 2021 Conference on Empirical Methods in Natural Language Processing","author":"Y Wang","year":"2021","unstructured":"Wang Y, Wang W, Joty S, Hoi SCH. CodeT5: identifier-aware unified pre-trained encoder-decoder models for code understanding and generation. In: Proceedings of 2021 Conference on Empirical Methods in Natural Language Processing. 2021, 8696\u20138708"},{"key":"50472_CR152","doi-asserted-by":"publisher","first-page":"1069","DOI":"10.18653\/v1\/2023.emnlp-main.68","volume-title":"Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing","author":"Y Wang","year":"2023","unstructured":"Wang Y, Le H, Gotmare AD, Bui NDQ, Li J, Hoi SCH. CodeT5+: open code large language models for code understanding and generation. In: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 1069\u20131088"},{"key":"50472_CR153","first-page":"1549","volume-title":"Proceedings of the 36th International Conference on Neural Information Processing Systems","author":"H Le","year":"2022","unstructured":"Le H, Wang Y, Gotmare AD, Savarese S, Hoi SCH. CodeRL: mastering code generation through pretrained models and deep reinforcement learning. In: Proceedings of the 36th International Conference on Neural Information Processing Systems. 2022, 1549"},{"key":"50472_CR154","unstructured":"Shojaee P, Jain A, Tipirneni S, Reddy CK. Execution-based code generation using deep reinforcement learning. Transactions on Machine Learning Research, 2023, 2023"},{"key":"50472_CR155","unstructured":"Ayupov S, Chirkova N. Parameter-efficient finetuning of transformers for source code. 2022, arXiv preprint arXiv: 2212.05901"},{"key":"50472_CR156","unstructured":"Zhuo TY, Zebaze A, Suppattarachai N, von Werra L, de Vries H, Liu Q, Muennighoff N. Astraios: parameter-efficient instruction tuning code large language models. 2024, arXiv preprint arXiv: 2401.00788"},{"issue":"7","key":"50472_CR157","doi-asserted-by":"publisher","first-page":"204","DOI":"10.1145\/3714461","volume":"34","author":"M Weyssow","year":"2025","unstructured":"Weyssow M, Zhou X, Kim K, Lo D, Sahraoui H. Exploring parameter-efficient fine-tuning techniques for code generation with large language models. ACM Transactions on Software Engineering and Methodology, 2025, 34(7): 204","journal-title":"ACM Transactions on Software Engineering and Methodology"},{"key":"50472_CR158","volume-title":"Proceedings of the 12th International Conference on Learning Representations","author":"H Xu","year":"2024","unstructured":"Xu H, Kim YJ, Sharaf A, Awadalla HH. A paradigm shift in machine translation: boosting translation performance of large language models. In: Proceedings of the 12th International Conference on Learning Representations. 2024"},{"key":"50472_CR159","volume-title":"Proceedings of the 41st International Conference on Machine Learning","author":"H Xu","year":"2024","unstructured":"Xu H, Sharaf A, Chen Y, Tan W, Shen L, Van Durme B, Murray K, Kim YJ. Contrastive preference optimization: pushing the boundaries of LLM performance in machine translation. In: Proceedings of the 41st International Conference on Machine Learning. 2024"},{"key":"50472_CR160","doi-asserted-by":"publisher","first-page":"482","DOI":"10.18653\/v1\/2023.wmt-1.44","volume-title":"Proceedings of the 8th Conference on Machine Translation","author":"V Iyer","year":"2023","unstructured":"Iyer V, Chen P, Birch A. Towards effective disambiguation for machine translation with large language models. In: Proceedings of the 8th Conference on Machine Translation. 2023, 482\u2013495"},{"key":"50472_CR161","unstructured":"Moslem Y, Haque R, Way A. Fine-tuning large language models for adaptive machine translation. 2023, arXiv preprint arXiv: 2312.12740"},{"key":"50472_CR162","doi-asserted-by":"publisher","first-page":"7919","DOI":"10.18653\/v1\/2022.emnlp-main.540","volume-title":"Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing","author":"A \u00dcst\u00fcn","year":"2022","unstructured":"\u00dcst\u00fcn A, Stickland AC. When does parameter-efficient transfer learning work for machine translation? In: Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing. 2022, 7919\u20137933"},{"key":"50472_CR163","doi-asserted-by":"publisher","first-page":"15432","DOI":"10.18653\/v1\/2023.findings-emnlp.1031","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023","author":"B Wu","year":"2023","unstructured":"Wu B, Yuan F, Zhao H, Li L, Xu J. Extrapolating multilingual understanding models as multilingual generators. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023. 2023, 15432\u201315444"},{"key":"50472_CR164","unstructured":"Wu M, Vu TT, Qu L, Foster G, Haffari G. Adapting large language models for document-level machine translation. 2024, arXiv preprint arXiv: 2401.06468"},{"key":"50472_CR165","volume-title":"Proceedings of the Thirteenth International Conference on Learning Representations","author":"H Luo","year":"2025","unstructured":"Luo H, Sun Q, Xu C, Zhao P, Lou JG, Tao C, Geng X, Lin Q, Chen S, Tang Y, Zhang D. WizardMath: empowering mathematical reasoning for large language models via reinforced evol-instruct. In: Proceedings of the Thirteenth International Conference on Learning Representations. 2025"},{"key":"50472_CR166","volume-title":"Proceedings of the 12th International Conference on Learning Representations","author":"X Yue","year":"2024","unstructured":"Yue X, Qu X, Zhang G, Fu Y, Huang W, Sun H, Su Y, Chen W. MammoTH: building math generalist models through hybrid instruction tuning. In: Proceedings of the 12th International Conference on Learning Representations. 2024"},{"key":"50472_CR167","doi-asserted-by":"publisher","first-page":"14852","DOI":"10.18653\/v1\/2023.acl-long.830","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"N Ho","year":"2023","unstructured":"Ho N, Schmid L, Yun SY. Large language models are reasoning teachers. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023, 14852\u201314882, doi: https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.830"},{"key":"50472_CR168","first-page":"2997","volume-title":"Proceedings of the 37th International Conference on Neural Information Processing Systems","author":"T Schick","year":"2023","unstructured":"Schick T, Dwivedi-Yu J, Dessi R, Raileanu R, Lomeli M, Hambro E, Zettlemoyer L, Cancedda N, Scialom T. Toolformer: language models can teach themselves to use tools. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. 2023, 2997"},{"key":"50472_CR169","doi-asserted-by":"publisher","first-page":"5254","DOI":"10.18653\/v1\/2023.emnlp-main.319","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"Z Hu","year":"2023","unstructured":"Hu Z, Wang L, Lan Y, Xu W, Lim EP, Bing L, Xu X, Poria S, Lee RKW. LLM-adapters: an adapter family for parameter-efficient fine-tuning of large language models. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 5254\u20135276"},{"key":"50472_CR170","doi-asserted-by":"publisher","first-page":"4663","DOI":"10.18653\/v1\/2024.findings-emnlp.268","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2024","author":"W Shi","year":"2024","unstructured":"Shi W, Hu Z, Bin Y, Liu J, Yang Y, Ng SK, Bing L, Lee RKW. Math-LLaVA: bootstrapping mathematical reasoning for multimodal large language models. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2024. 2024, 4663\u20134680"},{"key":"50472_CR171","unstructured":"Shao Z, Wang P, Zhu Q, Xu R, Song J, Bi X, Zhang H, Zhang M, Li YK, Wu Y, Guo D. DeepSeekMath: pushing the limits of mathematical reasoning in open language models. 2024, arXiv preprint arXiv: 2402.03300"},{"key":"50472_CR172","unstructured":"Luo L, Liu Y, Liu R, Phatale S, Guo M, Lara H, Li Y, Shu L, Zhu Y, Meng L, Sun J, Rastogi A. Improve mathematical reasoning in language models by automated process supervision. 2024, arXiv preprint arXiv: 2406.06592"},{"key":"50472_CR173","doi-asserted-by":"publisher","first-page":"5872","DOI":"10.18653\/v1\/2024.acl-long.320","volume-title":"Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"C Chen","year":"2024","unstructured":"Chen C, Wang X, Lin TE, Lv A, Wu Y, Gao X, Wen JR, Yan R, Li Y. Masked thought: simply masking partial reasoning steps can improve mathematical reasoning learning of language models. In: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2024, 5872\u20135900"},{"key":"50472_CR174","doi-asserted-by":"publisher","first-page":"2807","DOI":"10.18653\/v1\/2023.emnlp-main.169","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"T Liu","year":"2023","unstructured":"Liu T, Guo Q, Yang Y, Hu X, Zhang Y, Qiu X, Zhang Z. Plan, verify and switch: integrated reasoning with diverse X-of-thoughts. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 2807\u20132822"},{"key":"50472_CR175","unstructured":"Yu X, Zhou B, Cheng H, Roth D. ReasonAgain: using extractable symbolic programs to evaluate mathematical reasoning. 2024, arXiv preprint arXiv: 2410.19056"},{"key":"50472_CR176","doi-asserted-by":"publisher","first-page":"17222","DOI":"10.18653\/v1\/2025.acl-long.843","volume-title":"Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"L Ranaldi","year":"2025","unstructured":"Ranaldi L, Valentino M, Freitas A. Improving chain-of-thought reasoning via quasi-symbolic abstractions. In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2025, 17222\u201317240"},{"key":"50472_CR177","doi-asserted-by":"crossref","unstructured":"Srivastava G, Bi Z, Lu M, Wang X. DEBATE, TRAIN, EVOLVE: self evolution of language model reasoning. 2025, arXiv preprint arXiv: 2505.15734","DOI":"10.18653\/v1\/2025.emnlp-main.1666"},{"key":"50472_CR178","unstructured":"Cai H, Yang Y, Li Z. System-2 mathematical reasoning via enriched instruction tuning. 2024, arXiv preprint arXiv: 2412.16964"},{"issue":"6","key":"50472_CR179","doi-asserted-by":"publisher","first-page":"186357","DOI":"10.1007\/s11704-024-40555-y","volume":"18","author":"D Xu","year":"2024","unstructured":"Xu D, Chen W, Peng W, Zhang C, Xu T, Zhao X, Wu X, Zheng Y, Wang Y, Chen E. Large language models for generative information extraction: a survey. Frontiers of Computer Science, 2024, 18(6): 186357","journal-title":"Frontiers of Computer Science"},{"key":"50472_CR180","doi-asserted-by":"publisher","first-page":"100378","DOI":"10.1016\/j.health.2024.100378","volume":"7","author":"RM Siepmann","year":"2025","unstructured":"Siepmann RM, Baldini G, Schmidt CS, Truhn D, M\u00fcller-Franzes GA, Dada A, Kleesiek J, Nensa F, Hosch R. An automated information extraction model for unstructured discharge letters using large language models and GPT-4. Healthcare Analytics, 2025, 7: 100378","journal-title":"Healthcare Analytics"},{"issue":"1","key":"50472_CR181","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1186\/s12874-025-02470-z","volume":"25","author":"B Gu","year":"2025","unstructured":"Gu B, Shao V, Liao Z, Carducci V, Brufau SR, Yang J, Desai RJ. Scalable information extraction from free text electronic health records using large language models. BMC Medical Research Methodology, 2025, 25(1): 23","journal-title":"BMC Medical Research Methodology"},{"key":"50472_CR182","unstructured":"Xin Y, Luo S, Zhou H, Du J, Liu X, Fan Y, Li Q, Du Y. Parameter-efficient fine-tuning for pre-trained vision models: a survey. 2024, arXiv preprint arXiv: 2402.02242"},{"key":"50472_CR183","first-page":"2560","volume-title":"Proceedings of the 38th International Conference on Neural Information Processing Systems","author":"Y Xin","year":"2025","unstructured":"Xin Y, Luo S, Liu X, Du Y, Zhou H, Cheng X, Lee C, Du J, Wang H, Chen M, Liu T, Hu G, Wan Z, Zhang R, Li A, Yi M, Liu X. V-PETL bench: a unified visual parameter-efficient transfer learning benchmark. In: Proceedings of the 38th International Conference on Neural Information Processing Systems. 2025, 2560"},{"key":"50472_CR184","doi-asserted-by":"publisher","first-page":"4577","DOI":"10.24963\/ijcai.2021\/622","volume-title":"Proceedings of the 30th International Joint Conference on Artificial Intelligence","author":"L Qin","year":"2021","unstructured":"Qin L, Xie T, Che W, Liu T. A survey on spoken language understanding: recent advances and new frontiers. In: Proceedings of the 30th International Joint Conference on Artificial Intelligence., 2021, 4577\u20134584, doi: https:\/\/doi.org\/10.24963\/ijcai.2021\/622"},{"key":"50472_CR185","doi-asserted-by":"publisher","first-page":"391","DOI":"10.1109\/SLT.2016.7846294","volume-title":"Proceedings of 2016 IEEE Spoken Language Technology Workshop (SLT)","author":"R Sarikaya","year":"2016","unstructured":"Sarikaya R, Crook PA, Marin A, Jeong M, Robichaud JP, Celikyilmaz A, Kim YB, Rochette A, Khan OZ, Liu X, Boies D, Anastasakos T, Feizollahi Z, Ramesh N, Suzuki H, Holenstein R, Krawczyk E, Radostev V. An overview of end-to-end language understanding and dialog management for personal digital assistants. In: Proceedings of 2016 IEEE Spoken Language Technology Workshop (SLT). 2016, 391\u2013397"},{"key":"50472_CR186","doi-asserted-by":"publisher","first-page":"2428","DOI":"10.63317\/2ixggrdpuj7c","volume-title":"Proceedings of 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)","author":"Y Yoon","year":"2024","unstructured":"Yoon Y, Lee J, Kim K, Park C, Kim T. BlendX: complex multi-intent detection with blended patterns. In: Proceedings of 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024). 2024, 2428\u20132439"},{"key":"50472_CR187","first-page":"25038","volume-title":"Proceedings of the 39th AAAI Conference on Artificial Intelligence","author":"L Qin","year":"2025","unstructured":"Qin L, Chen Q, Zhou J, Wang J, Fei H, Che W, Li M. Divide-solve-combine: an interpretable and accurate prompting framework for zero-shot multi-intent detection. In: Proceedings of the 39th AAAI Conference on Artificial Intelligence. 2025, 25038\u201325046"},{"key":"50472_CR188","first-page":"1","volume-title":"Proceedings of 2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","author":"L Qin","year":"2025","unstructured":"Qin L, Wei F, Chen Q, Zhou J, Huang S, Si J, Lu W, Che W. CroPrompt: cross-task interactive prompting for zero-shot spoken language understanding. In: Proceedings of 2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 2025, 1\u20135"},{"key":"50472_CR189","first-page":"9147","volume-title":"Proceedings of the 31st International Conference on Computational Linguistics","author":"W Dong","year":"2025","unstructured":"Dong W, Chen S, Yang Y. ProTOD: proactive task-oriented dialogue system based on large language model. In: Proceedings of the 31st International Conference on Computational Linguistics. 2025, 9147\u20139164"},{"key":"50472_CR190","doi-asserted-by":"publisher","first-page":"12370","DOI":"10.18653\/v1\/2025.acl-long.605","volume-title":"Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"EC Acikgoz","year":"2025","unstructured":"Acikgoz EC, Greer J, Datta A, Yang Z, Zeng W, Elachqar O, Koukoumidis E, Hakkani-T\u00fcr D, Tur G. Can a single model master both multi-turn conversations and tool use? CALM: a unified conversational agentic language model. In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2025, 12370\u201312390"},{"key":"50472_CR191","first-page":"2616","volume-title":"Proceedings of the 31st International Conference on Computational Linguistics","author":"S Yin","year":"2025","unstructured":"Yin S, Huang P, Xu Y. MIDLM: multi-intent detection with bidirectional large language models. In: Proceedings of the 31st International Conference on Computational Linguistics. 2025, 2616\u20132625"},{"key":"50472_CR192","doi-asserted-by":"publisher","first-page":"174","DOI":"10.1007\/978-981-19-7596-7_14","volume-title":"Proceedings of the 7th China Conference on Knowledge Graph and Semantic Computing: Knowledge Graph Empowers the Digital Economy","author":"N Jin","year":"2022","unstructured":"Jin N, Siebert J, Li D, Chen Q. A survey on table question answering: recent advances. In: Proceedings of the 7th China Conference on Knowledge Graph and Semantic Computing: Knowledge Graph Empowers the Digital Economy. 2022, 174\u2013186"},{"key":"50472_CR193","unstructured":"Wang D, Dou L, Che W. A survey on table-and-text HybridQA: concepts, methods, challenges and future directions. 2022, arXiv preprint arXiv: 2212.13465"},{"issue":"9","key":"50472_CR194","doi-asserted-by":"publisher","first-page":"199348","DOI":"10.1007\/s11704-024-40330-z","volume":"19","author":"X Zhang","year":"2025","unstructured":"Zhang X, Wang D, Dou L, Zhu Q, Che W. A survey of table reasoning with large language models. Frontiers of Computer Science, 2025, 19(9): 199348","journal-title":"Frontiers of Computer Science"},{"key":"50472_CR195","doi-asserted-by":"crossref","unstructured":"Zhang X, Wang D, Xu K, Zhu Q, Che W. RoT: enhancing table reasoning with iterative row-wise traversals. 2025, arXiv preprint arXiv: 2505.15110","DOI":"10.18653\/v1\/2025.emnlp-main.29"},{"issue":"1","key":"50472_CR196","first-page":"1","volume":"2","author":"T Shi","year":"2018","unstructured":"Shi T, Keneshloo Y, Ramakrishnan N, Reddy CK. Neural abstractive text summarization with sequence-to-sequence models. ACM Transactions on Data Science, 2018, 2(1): 1","journal-title":"ACM Transactions on Data Science"},{"key":"50472_CR197","first-page":"208","volume-title":"Proceedings of the 1st International Conference on Business Intelligence, Computational Mathematics, and Data Analytics","author":"A Godbole","year":"2025","unstructured":"Godbole A, George JG, Shandilya S. Leveraging long-context large language models for multi-document understanding and summarization in enterprise applications. In: Proceedings of the 1st International Conference on Business Intelligence, Computational Mathematics, and Data Analytics. 2025, 208\u2013224"},{"issue":"4","key":"50472_CR198","doi-asserted-by":"publisher","first-page":"rsos241776","DOI":"10.1098\/rsos.241776","volume":"12","author":"U Peters","year":"2025","unstructured":"Peters U, Chin-Yee B. Generalization bias in large language model summarization of scientific research. Royal Society Open Science, 2025, 12(4): rsos241776","journal-title":"Royal Society Open Science"},{"key":"50472_CR199","first-page":"29724","volume-title":"Proceedings of the 39th AAAI Conference on Artificial Intelligence","author":"J Yun","year":"2025","unstructured":"Yun J, Choi J, Jin K, Jang S, Jang J, Kim Y. SummPilot: bridging efficiency and customization for interactive summarization system. In: Proceedings of the 39th AAAI Conference on Artificial Intelligence. 2025, 29724\u201329726"},{"key":"50472_CR200","first-page":"25065","volume-title":"Proceedings of the 39th AAAI Conference on Artificial Intelligence","author":"MR Qorib","year":"2025","unstructured":"Qorib MR, Hu Q, Ng HT. Just what you desire: constrained timeline summarization with self-reflection for enhanced relevance. In: Proceedings of the 39th AAAI Conference on Artificial Intelligence. 2025, 25065\u201325073"},{"key":"50472_CR201","unstructured":"Zhu DH, Xiong YJ, Zhang JC, Xie XJ, Xia CM. Understanding before reasoning: enhancing chain-of-thought with iterative summarization pre-prompting. 2025, arXiv preprint arXiv: 2501.04341"},{"key":"50472_CR202","first-page":"24930","volume-title":"Proceedings of the 39th AAAI Conference on Artificial Intelligence","author":"A Nandy","year":"2025","unstructured":"Nandy A, Bandyopadhyay S. Language models of code are few-shot planners and reasoners for multi-document summarization with attribution. In: Proceedings of the 39th AAAI Conference on Artificial Intelligence. 2025, 24930\u201324938"},{"key":"50472_CR203","doi-asserted-by":"publisher","first-page":"1368","DOI":"10.18653\/v1\/2023.acl-long.76","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Y Li","year":"2023","unstructured":"Li Y, Peng B, He P, Galley M, Yu Z, Gao J. DIONYSUS: a pre-trained model for low-resource dialogue summarization. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023, 1368\u20131386"},{"key":"50472_CR204","first-page":"25389","volume-title":"Proceedings of the 39th AAAI Conference on Artificial Intelligence","author":"L Wang","year":"2025","unstructured":"Wang L, Wu L, Song S, Wang Y, Gao C, Wang K. Distilling structured rationale from large language models to small language models for abstractive summarization. In: Proceedings of the 39th AAAI Conference on Artificial Intelligence. 2025, 25389\u201325397"},{"key":"50472_CR205","doi-asserted-by":"publisher","first-page":"7237","DOI":"10.18653\/v1\/2025.findings-naacl.404","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: NAACL 2025","author":"YJ Lu","year":"2025","unstructured":"Lu YJ, Hu TY, Koppula HS, Pouransari H, Chang JHR, Xia Y, Kong X, Zhu Q, Wang XS, Tuzel O, Vemulapalli R. Mutual reinforcement of LLM dialogue synthesis and summarization capabilities for few-shot dialogue summarization. In: Proceedings of Findings of the Association for Computational Linguistics: NAACL 2025. 2025, 7237\u20137256"},{"issue":"3","key":"50472_CR206","doi-asserted-by":"publisher","first-page":"470","DOI":"10.1093\/jamia\/ocae312","volume":"32","author":"A Aali","year":"2025","unstructured":"Aali A, Van Veen D, Arefeen YI, Hom J, Bluethgen C, Reis EP, Gatidis S, Clifford N, Daws J, Tehrani AS, Kim J, Chaudhari AS. A dataset and benchmark for hospital course summarization with adapted large language models. Journal of the American Medical Informatics Association, 2025, 32(3): 470\u2013479","journal-title":"Journal of the American Medical Informatics Association"},{"key":"50472_CR207","first-page":"25488","volume-title":"Proceedings of the 39th AAAI Conference on Artificial Intelligence","author":"J Wu","year":"2025","unstructured":"Wu J, Ning L, Liu L, Lee H, Wu N, Wang C, Prakash S, O\u2019Banion S, Green B, Xie J. RLPF: reinforcement learning from prediction feedback for user summarization with LLMs. In: Proceedings of the 39th AAAI Conference on Artificial Intelligence. 2025, 25488\u201325496"},{"key":"50472_CR208","unstructured":"Zhao H, Hui J, Howland J, Nguyen N, Zuo S, et al. CodeGemma: open code models based on gemma. 2024, arXiv preprint arXiv: 2406.11409"},{"key":"50472_CR209","unstructured":"Hui B, Yang J, Cui Z, Yang J, Liu D, et al. Qwen2.5-coder technical report. 2024, arXiv preprint arXiv: 2409.12186"},{"key":"50472_CR210","volume-title":"Seed-coder: let the code model curate data for itself","author":"B Dance-Seed","year":"2025","unstructured":"Dance-Seed B. Seed-coder: let the code model curate data for itself. See github.com\/ByteDance-Seed\/Seed-Coder\/blob\/master\/Seed-Coder website, 2025"},{"key":"50472_CR211","unstructured":"Cai W, Cao Y, Chen C, Chen C, Chen S, et al. Every sample matters: leveraging mixture-of-experts and high-quality data for efficient and accurate code LLM. 2025, arXiv preprint arXiv: 2503.17793"},{"key":"50472_CR212","volume-title":"Proceedings of the 1st Neural Information Processing Systems Track on Datasets and Benchmarks","author":"S Lu","year":"2021","unstructured":"Lu S, Guo D, Ren S, Huang J, Svyatkovskiy A, et al. CodeXGLUE: a machine learning benchmark dataset for code understanding and generation. In: Proceedings of the 1st Neural Information Processing Systems Track on Datasets and Benchmarks. 2021"},{"key":"50472_CR213","doi-asserted-by":"crossref","unstructured":"Ye Y, Zhang T, Jiang W, Huang H. Process-supervised reinforcement learning for code generation. 2025, arXiv preprint arXiv: 2502.01715","DOI":"10.18653\/v1\/2025.emnlp-main.719"},{"key":"50472_CR214","doi-asserted-by":"publisher","first-page":"9578","DOI":"10.18653\/v1\/2025.findings-acl.498","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: ACL 2025","author":"K Zhang","year":"2025","unstructured":"Zhang K, Li G, Li J, Dong Y, Jin Z. Focused-DPO: enhancing code generation through focused preference optimization on error-prone points. In: Proceedings of Findings of the Association for Computational Linguistics: ACL 2025. 2025, 9578\u20139591"},{"key":"50472_CR215","doi-asserted-by":"publisher","first-page":"12023","DOI":"10.18653\/v1\/2025.acl-long.587","volume-title":"Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"H Zeng","year":"2025","unstructured":"Zeng H, Jiang D, Wang H, Nie P, Chen X, Chen W. ACECODER: acing coder RL via automated test-case synthesis. In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2025, 12023\u201312040"},{"key":"50472_CR216","unstructured":"Wei Y, Duchenne O, Copet J, Carbonneaux Q, Zhang L, Fried D, Synnaeve G, Singh R, Wang SI. SWE-RL: advancing LLM reasoning via reinforcement learning on open software evolution. 2025, arXiv preprint arXiv: 2502.18449"},{"key":"50472_CR217","unstructured":"Storhaug A, Li J. Parameter-efficient fine-tuning of large language models for unit test generation: an empirical study. 2024, arXiv preprint arXiv: 2411.02462"},{"key":"50472_CR218","unstructured":"Zhang B, Liang P, Zhou X, Zhou X, Lo D, Feng Q, Li Z, Li L. A comprehensive evaluation of parameter-efficient fine-tuning on methodlevel code smell detection. 2024, arXiv preprint arXiv: 2412.13801"},{"key":"50472_CR219","volume-title":"Proceedings of 3rd International Conference on Learning Representations","author":"D Bahdanau","year":"2015","unstructured":"Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate. In: Proceedings of 3rd International Conference on Learning Representations. 2015"},{"key":"50472_CR220","doi-asserted-by":"publisher","first-page":"73","DOI":"10.1162\/tacl_a_00730","volume":"13","author":"J Pang","year":"2025","unstructured":"Pang J, Ye F, Wong DF, Yu D, Shi S, Tu Z, Wang L. Salute the classic: revisiting challenges of machine translation in the age of large language models. Transactions of the Association for Computational Linguistics, 2025, 13: 73\u201395","journal-title":"Transactions of the Association for Computational Linguistics"},{"key":"50472_CR221","doi-asserted-by":"publisher","first-page":"5028","DOI":"10.18653\/v1\/2024.emnlp-main.289","volume-title":"Proceedings of 2024 Conference on Empirical Methods in Natural Language Processing","author":"Y Huang","year":"2024","unstructured":"Huang Y, Li B, Feng X, Huo W, Fu C, Liu T, Qin B. Aligning translation-specific understanding to general understanding in large language models. In: Proceedings of 2024 Conference on Empirical Methods in Natural Language Processing. 2024, 5028\u20135041"},{"key":"50472_CR222","doi-asserted-by":"publisher","first-page":"16619","DOI":"10.63317\/2xcecragguex","volume-title":"Proceedings of 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)","author":"S Zhu","year":"2024","unstructured":"Zhu S, Cui M, Xiong D. Towards robust in-context learning for machine translation with large language models. In: Proceedings of 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024). 2024, 16619\u201316629"},{"key":"50472_CR223","doi-asserted-by":"crossref","unstructured":"Feng Z, Cao S, Ren J, Su J, Chen R, Zhang Y, Xu Z, Hu Y, Wu J, Liu Z. MT-R1-zero: advancing LLM-based machine translation via R1-zero-like reinforcement learning. 2025, arXiv preprint arXiv: 2504.10160","DOI":"10.18653\/v1\/2025.findings-emnlp.1015"},{"key":"50472_CR224","doi-asserted-by":"publisher","first-page":"15377","DOI":"10.18653\/v1\/2024.emnlp-main.860","volume-title":"Proceedings of 2024 Conference on Empirical Methods in Natural Language Processing","author":"Z Feng","year":"2024","unstructured":"Feng Z, Chen R, Zhang Y, Meng Z, Liu Z. Ladder: a model-agnostic framework boosting LLM-based machine translation to the next level. In: Proceedings of 2024 Conference on Empirical Methods in Natural Language Processing. 2024, 15377\u201315393"},{"key":"50472_CR225","doi-asserted-by":"publisher","first-page":"14605","DOI":"10.18653\/v1\/2023.acl-long.817","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"P Lu","year":"2023","unstructured":"Lu P, Qiu L, Yu W, Welleck S, Chang KW. A survey of deep learning for mathematical reasoning. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023, 14605\u201314631"},{"key":"50472_CR226","doi-asserted-by":"crossref","unstructured":"Yan Y, Wang S, Huo J, Yu PS, Hu X, Wen Q. Mathagent: leveraging a mixture-of-math-agent framework for real-world multimodal mathematical error detection. 2025, arXiv preprint arXiv: 2503.18132","DOI":"10.18653\/v1\/2025.acl-industry.7"},{"key":"50472_CR227","unstructured":"OpenAI. GPT-4 technical report. 2023, arXiv preprint arXiv: 2303.08774"},{"key":"50472_CR228","volume-title":"Elevating large language model reasoning ability with auto-enhanced zero-shot prompts","author":"Y Tang","year":"2025","unstructured":"Tang Y, Zhan Y, Zan C, Lan L, Che Y. Elevating large language model reasoning ability with auto-enhanced zero-shot prompts. Mathematical Foundations of Computing, 2025"},{"issue":"8055","key":"50472_CR229","doi-asserted-by":"publisher","first-page":"609","DOI":"10.1038\/s41586-025-08661-4","volume":"639","author":"M Yuksekgonul","year":"2025","unstructured":"Yuksekgonul M, Bianchi F, Boen J, Liu S, Lu P, Huang Z, Guestrin C, Zou J. Optimizing generative AI by backpropagating language model feedback. Nature, 2025, 639(8055): 609\u2013616","journal-title":"Nature"},{"key":"50472_CR230","doi-asserted-by":"crossref","unstructured":"Peng D, Zhou Y, Chen Q, Liu J, Chen J, Qin L. DLPO: towards a robust, efficient, and generalizable prompt optimization framework from a deep-learning perspective. 2025, arXiv preprint arXiv: 2503.13413","DOI":"10.18653\/v1\/2025.findings-emnlp.441"},{"key":"50472_CR231","unstructured":"Zhang B, Liu Y, Dong X, Zang Y, Zhang P, Duan H, Cao Y, Lin D, Wang J. BoostStep: boosting mathematical capability of large language models via improved single-step reasoning. 2025, arXiv preprint arXiv: 2501.03226"},{"key":"50472_CR232","unstructured":"Pang B, Dong H, Xu J, Savarese S, Zhou Y, Xiong C. BOLT: bootstrap long chain-of-thought in language models without distillation. 2025, arXiv preprint arXiv: 2502.03860"},{"key":"50472_CR233","doi-asserted-by":"publisher","first-page":"24914","DOI":"10.18653\/v1\/2025.acl-long.1213","volume-title":"Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Y Yu","year":"2025","unstructured":"Yu Y, Zhang Y, Zhang D, Liang X, Zhang H, Zhang X, Khademi M, Awadalla HH, Wang J, Yang Y, Wei F. Chain-of-reasoning: towards unified mathematical reasoning in large language models via a multi-paradigm perspective. In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2025, 24914\u201324937"},{"key":"50472_CR234","unstructured":"Qian C, Acikgoz EC, He Q, Wang H, Chen X, Hakkani-T\u00fcr D, Tur G, Ji H. ToolRL: reward is all tool learning needs. 2025, arXiv preprint arXiv: 2504.13958"},{"key":"50472_CR235","unstructured":"Singh J, Chakraborty T, Nambi A. Self-evolved preference optimization for enhancing mathematical reasoning in small language models. 2025, arXiv preprint arXiv: 2503.04813"},{"issue":"1","key":"50472_CR236","doi-asserted-by":"publisher","first-page":"30667","DOI":"10.1038\/s41598-024-75599-4","volume":"14","author":"NJ Prottasha","year":"2024","unstructured":"Prottasha NJ, Mahmud A, Sobuj MSI, Bhat P, Kowsher M, Yousefi N, Garibay OO. Parameter-efficient fine-tuning of large language models using semantic knowledge tuning. Scientific Reports, 2024, 14(1): 30667","journal-title":"Scientific Reports"},{"key":"50472_CR237","doi-asserted-by":"publisher","first-page":"7885","DOI":"10.18653\/v1\/2025.findings-naacl.440","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: NAACL 2025","author":"L Alazraki","year":"2025","unstructured":"Alazraki L, Rei M. Meta-reasoning improves tool use in large language models. In: Proceedings of Findings of the Association for Computational Linguistics: NAACL 2025. 2025, 7885\u20137897"},{"key":"50472_CR238","unstructured":"Qin L, Chen Q, Zhou Y, Chen Z, Li Y, Liao L, Li M, Che W, Yu PS. Multilingual large language model: a survey of resources, taxonomy and frontiers. 2024, arXiv preprint arXiv: 2404.04925"},{"key":"50472_CR239","doi-asserted-by":"publisher","first-page":"2936","DOI":"10.18653\/v1\/2023.findings-acl.185","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: ACL 2023","author":"G Winata","year":"2023","unstructured":"Winata G, Aji AF, Yong ZX, Solorio T. The decades progress on code-switching research in NLP: a systematic survey on trends and challenges. In: Proceedings of Findings of the Association for Computational Linguistics: ACL 2023. 2023, 2936\u20132978, doi: https:\/\/doi.org\/10.18653\/v1\/2023.findings-acl.185"},{"key":"50472_CR240","first-page":"28186","volume-title":"Proceedings of the 27th AAAI Conference on Artificial Intelligence","author":"Z Li","year":"2025","unstructured":"Li Z, Shi Y, Liu Z, Yang F, Payani A, Liu N, Du M. Language ranker: a metric for quantifying LLM performance across high and low-resource languages. In: Proceedings of the 27th AAAI Conference on Artificial Intelligence. 2025, 28186\u201328194"},{"key":"50472_CR241","doi-asserted-by":"publisher","first-page":"19320","DOI":"10.18653\/v1\/2025.findings-acl.988","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: ACL 2025","author":"P Wang","year":"2025","unstructured":"Wang P, Tao R, Chen Q, Hu M, Qin L. X-WebAgentBench: a multilingual interactive web benchmark for evaluating global agentic system. In: Proceedings of Findings of the Association for Computational Linguistics: ACL 2025. 2025, 19320\u201319335"},{"key":"50472_CR242","doi-asserted-by":"publisher","first-page":"30728","DOI":"10.18653\/v1\/2025.acl-long.1485","volume-title":"Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Y Zhang","year":"2025","unstructured":"Zhang Y, Liu X, Zhou R, Chen Q, Fei H, Lu W, Qin L. CCHaLL: a novel benchmark for joint cross-lingual and cross-modal hallucinations detection in large language models. In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2025, 30728\u201330749"},{"key":"50472_CR243","first-page":"483","volume-title":"Proceedings of 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","author":"L Xue","year":"2021","unstructured":"Xue L, Constant N, Roberts A, Kale M, Al-Rfou R, Siddhant A, Barua A, Raffel C. mt5: a massively multilingual pre-trained text-to-text transformer. In: Proceedings of 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2021, 483\u2013498"},{"key":"50472_CR244","unstructured":"Le Scao T, Fan A, Akiki C, Pavlick E, Ilic S, et al. BLOOM: a 176B-parameter open-access multilingual language model. 2022, arXiv preprint arXiv: 2211.05100"},{"key":"50472_CR245","doi-asserted-by":"publisher","first-page":"1347","DOI":"10.18653\/v1\/2024.findings-eacl.90","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EACL 2024","author":"P Chen","year":"2024","unstructured":"Chen P, Ji S, Bogoychev N, Kutuzov A, Haddow B, Heafield K. Monolingual or multilingual instruction tuning: which makes a better alpaca. In: Proceedings of Findings of the Association for Computational Linguistics: EACL 2024. 2024, 1347\u20131356"},{"key":"50472_CR246","unstructured":"Cahyawijaya S, Lovenia H, Yu T, Chung W, Fung P. Instruct-align: teaching novel languages with to LLMs through alignment-based cross-lingual instruction. 2023, arXiv preprint arXiv: 2305.13627"},{"key":"50472_CR247","doi-asserted-by":"publisher","first-page":"576","DOI":"10.1162\/tacl_a_00655","volume":"12","author":"J Li","year":"2024","unstructured":"Li J, Zhou H, Huang S, Cheng S, Chen J. Eliciting the translation ability of large language models via multilingual finetuning with translation instructions. Transactions of the Association for Computational Linguistics, 2024, 12: 576\u2013592","journal-title":"Transactions of the Association for Computational Linguistics"},{"key":"50472_CR248","first-page":"23469","volume-title":"Proceedings of the 39th AAAI Conference on Artificial Intelligence","author":"A Bajpai","year":"2025","unstructured":"Bajpai A, Chakraborty T. Multilingual LLMs inherently reward in-language time-sensitive semantic alignment for low-resource languages. In: Proceedings of the 39th AAAI Conference on Artificial Intelligence. 2025, 23469\u201323477"},{"key":"50472_CR249","first-page":"1","volume-title":"Proceedings of the 1st Workshop on Multilingual Representation Learning","author":"GI Winata","year":"2021","unstructured":"Winata GI, Madotto A, Lin Z, Liu R, Yosinski J, Fung P. Language models are few-shot multilingual learners. In: Proceedings of the 1st Workshop on Multilingual Representation Learning. 2021, 1\u201315"},{"key":"50472_CR250","volume-title":"Proceedings of the 11th International Conference on Learning Representations","author":"F Shi","year":"2023","unstructured":"Shi F, Suzgun M, Freitag M, Wang X, Srivats S, Vosoughi S, Chung HW, Tay Y, Ruder S, Zhou D, Das D, Wei J. Language models are multilingual chain-of-thought reasoners. In: Proceedings of the 11th International Conference on Learning Representations. 2023"},{"key":"50472_CR251","doi-asserted-by":"publisher","first-page":"9019","DOI":"10.18653\/v1\/2022.emnlp-main.616","volume-title":"Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing","author":"XV Lin","year":"2022","unstructured":"Lin XV, Mihaylov T, Artetxe M, Wang T, Chen S, Simig D, Ott M, Goyal N, Bhosale S, Du J, Pasunuru R, Shleifer S, Koura PS, Chaudhary V, O\u2019Horo B, Wang J, Zettlemoyer L, Kozareva Z, Diab M, Stoyanov V, Li X. Few-shot learning with multilingual generative language models. In: Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing. 2022, 9019\u20139052, doi: https:\/\/doi.org\/10.18653\/v1\/2022.emnlp-main.616"},{"key":"50472_CR252","doi-asserted-by":"publisher","first-page":"6292","DOI":"10.18653\/v1\/2023.acl-long.346","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"E Tanwar","year":"2023","unstructured":"Tanwar E, Dutta S, Borthakur M, Chakraborty T. Multilingual LLMs are better cross-lingual in-context learners with alignment. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023, 6292\u20136307, doi: https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.346"},{"key":"50472_CR253","doi-asserted-by":"publisher","first-page":"2695","DOI":"10.18653\/v1\/2023.emnlp-main.163","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"L Qin","year":"2023","unstructured":"Qin L, Chen Q, Wei F, Huang S, Che W. Cross-lingual prompting: improving zero-shot chain-of-thought reasoning across languages. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 2695\u20132709"},{"key":"50472_CR254","doi-asserted-by":"publisher","first-page":"12365","DOI":"10.18653\/v1\/2023.findings-emnlp.826","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023","author":"H Huang","year":"2023","unstructured":"Huang H, Tang T, Zhang D, Zhao X, Song T, Xia Y, Wei F. Not all languages are created equal in LLMs: improving multilingual capability by cross-lingual-thought prompting. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2023. 2023, 12365\u201312394, doi: https:\/\/doi.org\/10.18653\/v1\/2023.findings-emnlp.826"},{"issue":"6","key":"50472_CR255","doi-asserted-by":"publisher","first-page":"10675","DOI":"10.1109\/JIOT.2019.2940709","volume":"6","author":"Z Huang","year":"2019","unstructured":"Huang Z, Xu X, Ni J, Zhu H, Wang C. Multimodal representation learning for recommendation in internet of things. IEEE Internet of Things Journal, 2019, 6(6): 10675\u201310685","journal-title":"IEEE Internet of Things Journal"},{"key":"50472_CR256","unstructured":"Wang Y, Wu S, Zhang Y, Yan S, Liu Z, Luo J, Fei H. Multimodal chain-of-thought reasoning: a comprehensive survey. 2025, arXiv preprint arXiv: 2503.12605"},{"issue":"4","key":"50472_CR257","doi-asserted-by":"publisher","first-page":"2658","DOI":"10.1109\/TETCI.2025.3543389","volume":"9","author":"X Li","year":"2025","unstructured":"Li X, Qiao J, Yin S, Wu L, Gao C, Wang Z, Li X. A survey of multimodal fake news detection: a cross-modal interaction perspective. IEEE Transactions on Emerging Topics in Computational Intelligence, 2025, 9(4): 2658\u20132675","journal-title":"IEEE Transactions on Emerging Topics in Computational Intelligence"},{"key":"50472_CR258","unstructured":"Peng Y, Wang X, Wei Y, Pei J, Qiu W, Jian A, Hao Y, Pan J, Xie T, Ge L, Zhuang R, Song X, Liu Y, Zhou Y. Skywork R1V: pioneering multimodal reasoning with chain-of-thought. 2025, arXiv preprint arXiv: 2504.05599"},{"key":"50472_CR259","first-page":"182","volume-title":"Proceedings of the 36th International Conference on Neural Information Processing Systems","author":"P Lu","year":"2022","unstructured":"Lu P, Mishra S, Xia T, Qiu L, Chang KW, Zhu SC, Tafjord O, Clark P, Kalyan A. Learn to explain: multimodal reasoning via thought chains for science question answering. In: Proceedings of the 36th International Conference on Neural Information Processing Systems. 2022, 182"},{"key":"50472_CR260","doi-asserted-by":"publisher","first-page":"10834","DOI":"10.18653\/v1\/2023.findings-acl.689","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: ACL 2023","author":"L Qin","year":"2023","unstructured":"Qin L, Huang S, Chen Q, Cai C, Zhang Y, Liang B, Che W, Xu R. MMSD2.0: towards a reliable multi-modal sarcasm detection system. In: Proceedings of Findings of the Association for Computational Linguistics: ACL 2023. 2023, 10834\u201310845"},{"key":"50472_CR261","doi-asserted-by":"publisher","first-page":"1077","DOI":"10.18653\/v1\/2023.findings-acl.69","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: ACL 2023","author":"L Qin","year":"2023","unstructured":"Qin L, Wang W, Chen Q, Che W. CLIPText: a new paradigm for zero-shot text classification. In: Proceedings of Findings of the Association for Computational Linguistics: ACL 2023. 2023, 1077\u20131088"},{"key":"50472_CR262","unstructured":"Yang Z, Li L, Lin K, Wang J, Lin CC, Liu Z, Wang L. The dawn of LMMs: preliminary explorations with GPT-4V (ision). 2023, arXiv preprint arXiv: 2309.17421"},{"key":"50472_CR263","first-page":"13109","volume-title":"Proceedings of the 41st International Conference on Machine Learning","author":"H Fei","year":"2024","unstructured":"Fei H, Wu S, Ji W, Zhang H, Zhang M, Lee ML, Hsu W. Video-of-thought: step-by-step video reasoning from perception to cognition. In: Proceedings of the 41st International Conference on Machine Learning. 2024, 13109\u201313125"},{"key":"50472_CR264","volume-title":"Proceedings of the 38th International Conference on Neural Information Processing Systems","author":"L Qin","year":"2024","unstructured":"Qin L, Chen Q, Fei H, Chen Z, Li M, Che W. What factors affect multi-modal in-context learning? an in-depth exploration. In: Proceedings of the 38th International Conference on Neural Information Processing Systems. 2024"},{"key":"50472_CR265","doi-asserted-by":"crossref","unstructured":"Zhang Y, Liu X, Tao R, Chen Q, Fei H, Che W, Qin L. ViTCoT: video-text interleaved chain-of-thought for boosting video understanding in large language models. 2025, arXiv preprint arXiv: 2507.09876","DOI":"10.1145\/3746027.3755837"},{"key":"50472_CR266","volume-title":"Proceedings of the 38th International Conference on Neural Information Processing Systems","author":"W Wang","year":"2024","unstructured":"Wang W, Lv Q, Yu W, Hong W, Qi J, Wang Y, Ji J, Yang Z, Zhao L, Song X, Xu J, Xu B, Li J, Dong Y, Ding M, Tang J. CogVLM: visual expert for pretrained language models. In: Proceedings of the 38th International Conference on Neural Information Processing Systems. 2024"},{"key":"50472_CR267","doi-asserted-by":"publisher","first-page":"26286","DOI":"10.1109\/CVPR52733.2024.02484","volume-title":"Proceedings of 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","author":"H Liu","year":"2024","unstructured":"Liu H, Li C, Li Y, Lee YJ. Improved baselines with visual instruction tuning. In: Proceedings of 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2024, 26286\u201326296"},{"key":"50472_CR268","doi-asserted-by":"publisher","first-page":"8199","DOI":"10.18653\/v1\/2024.acl-long.446","volume-title":"Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Q Chen","year":"2024","unstructured":"Chen Q, Qin L, Zhang J, Chen Z, Xu X, Che W. M3CoT: a novel benchmark for multi-domain multi-step multi-modal chain-of-thought. In: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2024, 8199\u20138221"},{"key":"50472_CR269","unstructured":"Yang Z, Li L, Wang J, Lin K, Azarnasab E, Ahmed F, Liu Z, Liu C, Zeng M, Wang L. MM-REACT: prompting chatGPT for multimodal reasoning and action. 2023, arXiv preprint arXiv: 2303.11381"},{"key":"50472_CR270","volume-title":"Proceedings of the 12th International Conference on Learning Representations","author":"P Lu","year":"2024","unstructured":"Lu P, Bansal H, Xia T, Liu J, Li C, Hajishirzi H, Cheng H, Chang KW, Galley M, Gao J. MathVista: evaluating mathematical reasoning of foundation models in visual contexts. In: Proceedings of the 12th International Conference on Learning Representations. 2024"},{"key":"50472_CR271","unstructured":"Zhang Z, Zhang A, Li M, Zhao H, Karypis G, Smola A. Multimodal chain-of-thought reasoning in language models. Transactions on Machine Learning Research, 2024, 2024"},{"key":"50472_CR272","first-page":"23678","volume-title":"Proceedings of the 39th AAAI Conference on Artificial Intelligence","author":"Z Cheng","year":"2025","unstructured":"Cheng Z, Chen Q, Zhang J, Fei H, Feng X, Che W, Li M, Qin L. CoMT: a novel benchmark for chain of multi-modal thought on large vision-language models. In: Proceedings of the 39th AAAI Conference on Artificial Intelligence. 2025, 23678\u201323686"},{"key":"50472_CR273","unstructured":"Cheng Z, Chen Q, Xu X, Wang J, Wang W, Fei H, Wang Y, Wang AJ, Chen Z, Che W, Qin L. Visual thoughts: a unified perspective of understanding multimodal chain-of-thought. 2025, arXiv preprint arXiv: 2505.15510"},{"key":"50472_CR274","unstructured":"Wu Y, Zhang P, Xiong W, Oguz B, Gee JC, Nie Y. The role of chain-of-thought in complex vision-language reasoning task. 2023, arXiv preprint arXiv: 2311.09193"},{"key":"50472_CR275","doi-asserted-by":"publisher","first-page":"14420","DOI":"10.1109\/CVPR52733.2024.01367","volume-title":"Proceedings of 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","author":"C Mitra","year":"2024","unstructured":"Mitra C, Huang B, Darrell T, Herzig R. Compositional chain-of-thought prompting for large multimodal models. In: Proceedings of 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2024, 14420\u201314431"},{"key":"50472_CR276","doi-asserted-by":"crossref","unstructured":"Wang P, Zhang Y, Fei H, Chen Q, Wang Y, Si J, Lu W, Li M, Qin L. S3 agent: unlocking the power of VLLM for zero-shot multimodal sarcasm detection. ACM Transactions on Multimedia Computing, Communications and Applications, 2024","DOI":"10.1145\/3690642"},{"key":"50472_CR277","volume-title":"Proceedings of the 12th International Conference on Learning Representations","author":"Y Qin","year":"2024","unstructured":"Qin Y, Liang S, Ye Y, Zhu K, Yan L, Lu Y, Lin Y, Cong X, Tang X, Qian B, Zhao S, Hong L, Tian R, Xie R, Zhou J, Gerstein M, Li D, Liu Z, Sun M. ToolLLM: facilitating large language models to master 16000+ real-world APIs. In: Proceedings of the 12th International Conference on Learning Representations. 2024"},{"key":"50472_CR278","volume-title":"Proceedings of the 12th International Conference on Learning Representations","author":"M Hu","year":"2024","unstructured":"Hu M, Mu Y, Yu XC, Ding M, Wu S, Shao W, Chen Q, Wang B, Qiao Y, Luo P. Tree-planner: efficient close-loop task planning with large language models. In: Proceedings of the 12th International Conference on Learning Representations. 2024"},{"key":"50472_CR279","first-page":"377","volume-title":"Proceedings of the 37th International Conference on Neural Information Processing Systems","author":"N Shinn","year":"2023","unstructured":"Shinn N, Cassano F, Gopinath A, Narasimhan KR, Yao S. Reflexion: language agents with verbal reinforcement learning. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. 2023, 377"},{"issue":"6","key":"50472_CR280","doi-asserted-by":"publisher","first-page":"186345","DOI":"10.1007\/s11704-024-40231-1","volume":"18","author":"L Wang","year":"2024","unstructured":"Wang L, Ma C, Feng X, Zhang Z, Yang H, Zhang J, Chen Z, Tang J, Chen X, Lin Y, Zhao WX, Wei Z, Wen J. A survey on large language model based autonomous agents. Frontiers of Computer Science, 2024, 18(6): 186345","journal-title":"Frontiers of Computer Science"},{"key":"50472_CR281","unstructured":"Zhu X, Chen Y, Tian H, Tao C, Su W, Yang C, Huang G, Li B, Lu L, Wang X, Qiao Y, Zhang Z, Dai J. Ghost in the minecraft: generally capable agents for open-world environments via large language models with text-based knowledge and memory. 2023, arXiv preprint arXiv: 2305.17144"},{"key":"50472_CR282","doi-asserted-by":"publisher","first-page":"32779","DOI":"10.18653\/v1\/2025.acl-long.1575","volume-title":"Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"M Hu","year":"2025","unstructured":"Hu M, Chen T, Chen Q, Mu Y, Shao W, Luo P. HiAgent: hierarchical working memory management for solving long-horizon agent tasks with large language model. In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2025, 32779\u201332798"},{"key":"50472_CR283","unstructured":"Zhang G, Niu L, Fang J, Wang K, Bai L, Wang X. Multi-agent architecture search via agentic supernet. 2025, arXiv preprint arXiv: 2502.04180"},{"key":"50472_CR284","doi-asserted-by":"publisher","first-page":"15549","DOI":"10.18653\/v1\/2025.acl-long.757","volume-title":"Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Y Yue","year":"2025","unstructured":"Yue Y, Zhang G, Liu B, Wan G, Wang K, Cheng D, Qi Y. MasRouter: learning to route LLMs for multi-agent systems. In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2025, 15549\u201315572"},{"key":"50472_CR285","first-page":"517","volume-title":"Proceedings of the 37th International Conference on Neural Information Processing Systems","author":"S Yao","year":"2023","unstructured":"Yao S, Yu D, Zhao J, Shafran I, Griffiths TL, Cao Y, Narasimhan K. Tree of thoughts: deliberate problem solving with large language models. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. 2023, 517"},{"key":"50472_CR286","unstructured":"Chen W, Ma X, Wang X, Cohen WW. Program of thoughts prompting: disentangling computation from reasoning for numerical reasoning tasks. Transactions on Machine Learning Research, 2023, 2023"},{"key":"50472_CR287","unstructured":"Lei B, Lin PH, Liao C, Ding C. Boosting logical reasoning in large language models through a new framework: the graph of thought. 2023, arXiv preprint arXiv: 2308.08614"},{"key":"50472_CR288","doi-asserted-by":"publisher","first-page":"6644","DOI":"10.18653\/v1\/2024.findings-emnlp.388","volume-title":"Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2024","author":"Y Zhang","year":"2024","unstructured":"Zhang Y, Chen Q, Zhou J, Wang P, Si J, Wang J, Lu W, Qin L. Wrong-of-thought: an integrated reasoning framework with multi-perspective verification and wrong information. In: Proceedings of Findings of the Association for Computational Linguistics: EMNLP 2024. 2024, 6644\u20136653"},{"key":"50472_CR289","first-page":"49","volume-title":"Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"D Muhlgay","year":"2024","unstructured":"Muhlgay D, Ram O, Magar I, Levine Y, Ratner N, Belinkov Y, Abend O, Leyton-Brown K, Shashua A, Shoham Y. Generating benchmarks for factuality evaluation of language models. In: Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers). 2024, 49\u201366"},{"key":"50472_CR290","doi-asserted-by":"publisher","first-page":"12076","DOI":"10.18653\/v1\/2023.emnlp-main.741","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"S Min","year":"2023","unstructured":"Min S, Krishna K, Lyu X, Lewis M, Yih WT, Koh PW, Iyyer M, Zettlemoyer L, Hajishirzi H. FActScore: fine-grained atomic evaluation of factual precision in long form text generation. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 12076\u201312100"},{"key":"50472_CR291","doi-asserted-by":"publisher","first-page":"681","DOI":"10.1162\/tacl_a_00667","volume":"12","author":"V Adlakha","year":"2024","unstructured":"Adlakha V, BehnamGhader P, Lu XH, Meade N, Reddy S. Evaluating correctness and faithfulness of instruction-following models for question answering. Transactions of the Association for Computational Linguistics, 2024, 12: 681\u2013699","journal-title":"Transactions of the Association for Computational Linguistics"},{"key":"50472_CR292","doi-asserted-by":"publisher","first-page":"6723","DOI":"10.18653\/v1\/2022.acl-long.464","volume-title":"Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"T Liu","year":"2022","unstructured":"Liu T, Zhang Y, Brockett C, Mao Y, Sui Z, Chen W, Dolan WB. A token-level reference-free hallucination detection benchmark for freeform text generation. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2022, 6723\u20136737"},{"key":"50472_CR293","doi-asserted-by":"publisher","first-page":"7312","DOI":"10.18653\/v1\/2023.emnlp-main.453","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"KK Chang","year":"2023","unstructured":"Chang KK, Cramer M, Soni S, Bamman D. Speak, memory: an archaeology of books known to ChatGPT\/GPT-4. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 7312\u20137327"},{"key":"50472_CR294","doi-asserted-by":"publisher","first-page":"3309","DOI":"10.18653\/v1\/2022.acl-long.234","volume-title":"Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"T Hartvigsen","year":"2022","unstructured":"Hartvigsen T, Gabriel S, Palangi H, Sap M, Ray D, Kamar E. ToxiGen: a large-scale machine-generated dataset for adversarial and implicit hate speech detection. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2022, 3309\u20133326, doi: https:\/\/doi.org\/10.18653\/v1\/2022.acl-long.234"},{"key":"50472_CR295","doi-asserted-by":"publisher","first-page":"515","DOI":"10.1145\/3611643.3616310","volume-title":"Proceedings of the 31st ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering","author":"Y Wan","year":"2023","unstructured":"Wan Y, Wang W, He P, Gu J, Bai H, Lyu MR. BiasAsker: measuring the bias in conversational AI system. In: Proceedings of the 31st ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering. 2023, 515\u2013527"},{"key":"50472_CR296","doi-asserted-by":"publisher","first-page":"862","DOI":"10.1145\/3442188.3445924","volume-title":"Proceedings of 2021 ACM Conference on Fairness, Accountability, and Transparency","author":"J Dhamala","year":"2021","unstructured":"Dhamala J, Sun T, Kumar V, Krishna S, Pruksachatkun Y, Chang KW, Gupta R. BOLD: dataset and metrics for measuring biases in open-ended language generation. In: Proceedings of 2021 ACM Conference on Fairness, Accountability, and Transparency. 2021, 862\u2013872"},{"key":"50472_CR297","unstructured":"Ganguli D, Lovitt L, Kernion J, Askell A, Bai Y, et al. Red teaming language models to reduce harms: methods, scaling behaviors, and lessons learned. 2022, arXiv preprint arXiv: 2209.07858"},{"key":"50472_CR298","unstructured":"Sun H, Zhang Z, Deng J, Cheng J, Huang M. Safety assessment of Chinese large language models. 2023, arXiv preprint arXiv: 2304.10436"},{"key":"50472_CR299","unstructured":"Pan W, Liu Z, Chen Q, Zhou X, Yu H, Jia X. The hidden dimensions of LLM alignment: a multi-dimensional safety analysis. 2025, arXiv preprint arXiv: 2502.09674"},{"key":"50472_CR300","doi-asserted-by":"publisher","first-page":"6216","DOI":"10.1145\/3711896.3736561","volume-title":"Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V.2","author":"M Yu","year":"2025","unstructured":"Yu M, Meng F, Zhou X, Wang S, Mao J, Pan L, Chen T, Wang K, Li X, Zhang Y, An B, Wen Q. A survey on trustworthy LLM agents: threats and countermeasures. In: Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V.2. 2025, 6216\u20136226"},{"issue":"11","key":"50472_CR301","doi-asserted-by":"publisher","first-page":"1911362","DOI":"10.1007\/s11704-024-40579-4","volume":"19","author":"Y Xu","year":"2025","unstructured":"Xu Y, Hu L, Zhao J, Qiu Z, Xu K, Ye Y, Gu H. A survey on multilingual large language models: corpora, alignment, and bias. Frontiers of Computer Science, 2025, 19(11): 1911362","journal-title":"Frontiers of Computer Science"},{"key":"50472_CR302","unstructured":"Li ZZ, Zhang D, Zhang ML, Zhang J, Liu Z, et al. From system 1 to system 2: a survey of reasoning large language models. 2025, arXiv preprint arXiv: 2502.17419"},{"key":"50472_CR303","unstructured":"Jaech A, Kalai A, Lerer A, Richardson A, El-Kishky A, et al. OpenAI o1 system card. 2024, arXiv preprint arXiv: 2412.16720"},{"key":"50472_CR304","doi-asserted-by":"publisher","first-page":"2665","DOI":"10.18653\/v1\/2023.acl-long.150","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"LH Li","year":"2023","unstructured":"Li LH, Hessel J, Yu Y, Ren X, Chang KW, Choi Y. Symbolic chain-of-thought distillation: small models can also \u201cthink\u201d step-by-step. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023, 2665\u20132679, doi: https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.150"},{"key":"50472_CR305","doi-asserted-by":"publisher","first-page":"5546","DOI":"10.18653\/v1\/2023.acl-long.304","volume-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"P Wang","year":"2023","unstructured":"Wang P, Wang Z, Li Z, Gao Y, Yin B, Ren X. SCOTT: self-consistent chain-of-thought distillation. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023, 5546\u20135558"},{"key":"50472_CR306","unstructured":"Chen Q, Qin L, Liu J, Peng D, Wang J, Hu M, Chen Z, Che W, Liu T. ECM: a unified electronic circuit model for explaining the emergence of in-context learning and chain-of-thought in large language model. 2025, arXiv preprint arXiv: 2502.03325"},{"key":"50472_CR307","first-page":"305","volume-title":"Proceedings of the 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics","author":"Q Lyu","year":"2023","unstructured":"Lyu Q, Havaldar S, Stein A, Zhang L, Rao D, Wong E, Apidianaki M, Callison-Burch C. Faithful chain-of-thought reasoning. In: Proceedings of the 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics. 2023, 305\u2013329"},{"key":"50472_CR308","unstructured":"Zeng S, Chang X, Xie M, Liu X, Bai Y, Pan Z, Xu M, Wei X. FutureSightDrive: thinking visually with spatio-temporal cot for autonomous driving. 2025, arXiv preprint arXiv: 2505.17685"},{"key":"50472_CR309","unstructured":"Renze M, Guven E. Self-reflection in LLM agents: effects on problem-solving performance. 2024, arXiv preprint arXiv: 2405.06682"},{"key":"50472_CR310","unstructured":"Balachandran V, Chen J, Chen L, Garg S, Joshi N, Lara Y, Langford J, Nushi B, Vineet V, Wu Y, Yousefi S. Inference-time scaling for complex tasks: where we stand and what lies ahead. 2025, arXiv preprint arXiv: 2504.00294"},{"key":"50472_CR311","unstructured":"Wu Y, Sun Z, Li S, Welleck S, Yang Y. Inference scaling laws: an empirical analysis of compute-optimal inference for problem-solving with language models. 2024, arXiv preprint arXiv: 2408.00724"},{"key":"50472_CR312","unstructured":"Yu Q, Zhang Z, Zhu R, Yuan Y, Zuo X, et al. DAPO: an open-source llm reinforcement learning system at scale. 2025, arXiv preprint arXiv: 2503.14476"},{"key":"50472_CR313","unstructured":"Yue Y, Yuan Y, Yu Q, Zuo X, Zhu R, et al. VAPO: efficient and reliable reinforcement learning for advanced reasoning tasks. 2025, arXiv preprint arXiv: 2504.05118"},{"key":"50472_CR314","unstructured":"Chen J, Fan T, Liu X, Liu L, Lin Z, et al. Seed-thinking-v1.5: advancing superb reasoning models with reinforcement learning. 2025, arXiv preprint arXiv: 2504.13914"},{"key":"50472_CR315","unstructured":"Duan K, Liu Z, Mao X, Pang T, Chen C, Chen Q, Shieh MQ, Dou L. Efficient process reward model training via active learning. 2025, arXiv preprint arXiv: 2504.10559"},{"key":"50472_CR316","unstructured":"Sui Y, Chuang YN, Wang G, Zhang J, Zhang T, Yuan J, Liu H, Wen A, Zhong S, Zou N, Chen H, Hu X. Stop overthinking: a survey on efficient reasoning for large language models. Transactions on Machine Learning Research, 2025, 2025"},{"key":"50472_CR317","unstructured":"Feng S, Fang G, Ma X, Wang X. Efficient reasoning models: a survey. 2025, arXiv preprint arXiv: 2504.10903"},{"key":"50472_CR318","unstructured":"Hou B, Zhang Y, Ji J, Liu Y, Qian K, Andreas J, Chang S. ThinkPrune: pruning long chain-of-thought of llms via reinforcement learning. 2025, arXiv preprint arXiv: 2504.01296"},{"key":"50472_CR319","unstructured":"Chen Q, Qin L, Liu J, Liao Y, Wang J, Zhou J, Che W. RBF++: quantifying and optimizing reasoning boundaries across measurable and unmeasurable capabilities for chain-of-thought reasoning. 2025, arXiv preprint arXiv: 2505.13307"},{"key":"50472_CR320","doi-asserted-by":"publisher","first-page":"4177","DOI":"10.18653\/v1\/2023.emnlp-main.255","volume-title":"Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing","author":"J Qi","year":"2023","unstructured":"Qi J, Xu Z, Shen Y, Liu M, Jin D, Wang Q, Huang L. The art of SOCRATIC QUESTIONING: recursive thinking with large language models. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 4177\u20134199"},{"key":"50472_CR321","first-page":"1100","volume-title":"Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"D Paul","year":"2024","unstructured":"Paul D, Ismayilzada M, Peyrard M, Borges B, Bosselut A, West R, Faltings B. REFINER: reasoning feedback on intermediate representations. In: Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers). 2024, 1100\u20131126"},{"key":"50472_CR322","first-page":"2019","volume-title":"Proceedings of the 37th International Conference on Neural Information Processing Systems","author":"A Madaan","year":"2023","unstructured":"Madaan A, Tandon N, Gupta P, Hallinan S, Gao L, Wiegreffe S, Alon U, Dziri N, Prabhumoye S, Yang Y, Gupta S, Majumder BP, Hermann K, Welleck S, Yazdanbakhsh A, Clark P. SELF-REFINE: iterative refinement with self-feedback. In: Proceedings of the 37th International Conference on Neural Information Processing Systems. 2023, 2019"},{"key":"50472_CR323","unstructured":"Li Y, Shen X, Yao X, Ding X, Miao Y, Krishnan R, Padman R. Beyond single-turn: a survey on multi-turn interactions with large language models. 2025, arXiv preprint arXiv: 2504.04717"},{"key":"50472_CR324","volume-title":"Proceedings of the 11th International Conference on Learning Representations","author":"S Yao","year":"2023","unstructured":"Yao S, Zhao J, Yu D, Du N, Shafran I, Narasimhan KR, Cao Y. ReAct: synergizing reasoning and acting in language models. In: Proceedings of the 11th International Conference on Learning Representations. 2023"},{"key":"50472_CR325","doi-asserted-by":"publisher","first-page":"27129","DOI":"10.18653\/v1\/2025.acl-long.1316","volume-title":"Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","author":"Z Chen","year":"2025","unstructured":"Chen Z, Chen Q, Qin L, Guo Q, Lv H, Zou Y, Yan H, Chen K, Lin D. What are the essential factors in crafting effective long context multi-hop instruction datasets? Insights and best practices. In: Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2025, 27129\u201327151"}],"container-title":["Frontiers of Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11704-025-50472-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11704-025-50472-3","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11704-025-50472-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T06:04:04Z","timestamp":1772863444000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11704-025-50472-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,7]]},"references-count":325,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2026,11]]}},"alternative-id":["50472"],"URL":"https:\/\/doi.org\/10.1007\/s11704-025-50472-3","relation":{},"ISSN":["2095-2228","2095-2236"],"issn-type":[{"value":"2095-2228","type":"print"},{"value":"2095-2236","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,3,7]]},"assertion":[{"value":"17 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 August 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 March 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors declare that they have no competing interests or financialconflicts to disclose.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"2011361"}}