{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T14:27:05Z","timestamp":1776090425991,"version":"3.50.1"},"reference-count":66,"publisher":"Association for Computing Machinery (ACM)","issue":"8","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. VLDB Endow."],"published-print":{"date-parts":[[2024,4]]},"abstract":"<jats:p>Table Question Answering (TQA) presents a substantial challenge at the intersection of natural language processing and data analytics. This task involves answering natural language (NL) questions on top of tabular data, demanding proficiency in logical reasoning, understanding of data semantics, and fundamental analytical capabilities. Due to its significance, a substantial volume of research has been dedicated to exploring a wide range of strategies aimed at tackling this challenge including approaches that leverage Large Language Models (LLMs) through in-context learning or Chain-of-Thought (CoT) prompting as well as approaches that train and fine-tune custom models.<\/jats:p>\n          <jats:p>\n            Nonetheless, a conspicuous gap exists in the research landscape, where there is limited exploration of how innovative foundational research, which integrates incremental reasoning with external tools in the context of LLMs, as exemplified by the ReAct paradigm, could potentially bring advantages to the TQA task. In this paper, we aim to fill this gap, by introducing ReAcTable (\n            <jats:bold>ReAct<\/jats:bold>\n            for\n            <jats:bold>Table<\/jats:bold>\n            Question Answering tasks), a framework inspired by the ReAct paradigm that is carefully enhanced to address the challenges uniquely appearing in TQA tasks such as interpreting complex data semantics, dealing with errors generated by inconsistent data and generating intricate data transformations. ReAcTable relies on external tools such as SQL and Python code executors, to progressively enhance the data by generating intermediate data representations, ultimately transforming it into a more accessible format for answering the user's questions with greater ease. Through extensive empirical evaluations using three popular TQA benchmarks, we demonstrate that ReAcTable achieves remarkable performance even when compared to fine-tuned approaches. In particular, it outperforms the best prior result on the WikiTQ benchmark by 2.1%, achieving an accuracy of 68.0% without requiring training a new model or fine-tuning.\n          <\/jats:p>","DOI":"10.14778\/3659437.3659452","type":"journal-article","created":{"date-parts":[[2024,5,31]],"date-time":"2024-05-31T16:22:27Z","timestamp":1717172547000},"page":"1981-1994","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":35,"title":["ReAcTable: Enhancing ReAct for Table Question Answering"],"prefix":"10.14778","volume":"17","author":[{"given":"Yunjia","family":"Zhang","sequence":"first","affiliation":[{"name":"University of Wisconsin-Madison"}]},{"given":"Jordan","family":"Henkel","sequence":"additional","affiliation":[{"name":"Microsoft"}]},{"given":"Avrilia","family":"Floratou","sequence":"additional","affiliation":[{"name":"Microsoft"}]},{"given":"Joyce","family":"Cahoon","sequence":"additional","affiliation":[{"name":"Microsoft"}]},{"given":"Shaleen","family":"Deep","sequence":"additional","affiliation":[{"name":"Microsoft"}]},{"given":"Jignesh M.","family":"Patel","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University"}]}],"member":"320","published-online":{"date-parts":[[2024,5,31]]},"reference":[{"key":"e_1_2_1_1_1","unstructured":"Tom B. Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell Sandhini Agarwal Ariel Herbert-Voss Gretchen Krueger Tom Henighan Rewon Child Aditya Ramesh Daniel M. Ziegler Jeffrey Wu Clemens Winter Christopher Hesse Mark Chen Eric Sigler Mateusz Litwin Scott Gray Benjamin Chess Jack Clark Christopher Berner Sam McCandlish Alec Radford Ilya Sutskever and Dario Amodei. 2020. Language Models are Few-Shot Learners. arXiv:2005.14165 [cs.CL]"},{"key":"e_1_2_1_2_1","unstructured":"Mark Chen Jerry Tworek Heewoo Jun Qiming Yuan Henrique Ponde de Oliveira Pinto Jared Kaplan Harri Edwards Yuri Burda Nicholas Joseph Greg Brockman Alex Ray Raul Puri Gretchen Krueger Michael Petrov Heidy Khlaaf Girish Sastry Pamela Mishkin Brooke Chan Scott Gray Nick Ryder Mikhail Pavlov Alethea Power Lukasz Kaiser Mohammad Bavarian Clemens Winter Philippe Tillet Felipe Petroski Such Dave Cummings Matthias Plappert Fotios Chantzis Elizabeth Barnes Ariel Herbert-Voss William Hebgen Guss Alex Nichol Alex Paino Nikolas Tezak Jie Tang Igor Babuschkin Suchir Balaji Shantanu Jain William Saunders Christopher Hesse Andrew N. Carr Jan Leike Josh Achiam Vedant Misra Evan Morikawa Alec Radford Matthew Knight Miles Brundage Mira Murati Katie Mayer Peter Welinder Bob McGrew Dario Amodei Sam McCandlish Ilya Sutskever and Wojciech Zaremba. 2021. Evaluating Large Language Models Trained on Code. arXiv:2107.03374 [cs.LG]"},{"key":"e_1_2_1_3_1","volume-title":"TabFact: A Large-scale Dataset for Table-based Fact Verification. CoRR abs\/1909.02164","author":"Chen Wenhu","year":"2019","unstructured":"Wenhu Chen, Hongmin Wang, Jianshu Chen, Yunkai Zhang, Hong Wang, Shiyang Li, Xiyou Zhou, and William Yang Wang. 2019. TabFact: A Large-scale Dataset for Table-based Fact Verification. CoRR abs\/1909.02164 (2019). arXiv:1909.02164 http:\/\/arxiv.org\/abs\/1909.02164"},{"key":"e_1_2_1_4_1","volume-title":"TabFact: A Large-scale Dataset for Table-based Fact Verification. International Conference on Learning Representations","author":"Chen Wenhu","year":"2020","unstructured":"Wenhu Chen, Hongmin Wang, Jianshu Chen, Yunkai Zhang, Hong Wang, Shiyang Li, Xiyou Zhou, and William Yang Wang. 2020. TabFact: A Large-scale Dataset for Table-based Fact Verification. International Conference on Learning Representations (2020). https:\/\/openreview.net\/forum?id=rkeJRhNYDH"},{"key":"e_1_2_1_5_1","unstructured":"Zhoujun Cheng Tianbao Xie Peng Shi Chengzu Li Rahul Nadkarni Yushi Hu Caiming Xiong Dragomir Radev Mari Ostendorf Luke Zettlemoyer Noah A. Smith and Tao Yu. 2023. Binding Language Models in Symbolic Languages. arXiv:2210.02875 [cs.CL]"},{"key":"e_1_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1273"},{"key":"e_1_2_1_7_1","volume-title":"Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:2002.08155","author":"Feng Zhangyin","year":"2020","unstructured":"Zhangyin Feng, Daya Guo, Duyu Tang, Nan Duan, Xiaocheng Feng, Ming Gong, Linjun Shou, Bing Qin, Ting Liu, Daxin Jiang, et al. 2020. Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:2002.08155 (2020)."},{"key":"e_1_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3180155.3180167"},{"key":"e_1_2_1_9_1","volume-title":"PASTA: Table-Operations Aware Fact Verification via Sentence-Table Cloze Pre-training. arXiv:2211.02816 [cs.CL]","author":"Gu Zihui","year":"2022","unstructured":"Zihui Gu, Ju Fan, Nan Tang, Preslav Nakov, Xiaoman Zhao, and Xiaoyong Du. 2022. PASTA: Table-Operations Aware Fact Verification via Sentence-Table Cloze Pre-training. arXiv:2211.02816 [cs.CL]"},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/2814270.2814295"},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.398"},{"key":"e_1_2_1_12_1","unstructured":"Richard D Hipp. 2020. SQLite. https:\/\/www.sqlite.org\/index.html"},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2005.1553554"},{"key":"e_1_2_1_14_1","doi-asserted-by":"crossref","unstructured":"Zhengbao Jiang Yi Mao Pengcheng He Graham Neubig and Weizhu Chen. 2022. OmniTab: Pretraining with Natural and Synthetic Data for Few-shot Table-based Question Answering. arXiv:2207.03637 [cs.CL]","DOI":"10.18653\/v1\/2022.naacl-main.68"},{"key":"e_1_2_1_15_1","unstructured":"Nengzheng Jin Joanna Siebert Dongfang Li and Qingcai Chen. 2022. A Survey on Table Question Answering: Recent Advances. arXiv:2207.05270 [cs.CL]"},{"key":"e_1_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3555041.3589678"},{"key":"e_1_2_1_17_1","volume-title":"Evolutionary Reinforcement Learning for Sample-Efficient Multiagent Coordination. CoRR abs\/1906.07315","author":"Khadka Shauharda","year":"2019","unstructured":"Shauharda Khadka, Somdeb Majumdar, and Kagan Tumer. 2019. Evolutionary Reinforcement Learning for Sample-Efficient Multiagent Coordination. CoRR abs\/1906.07315 (2019). arXiv:1906.07315 http:\/\/arxiv.org\/abs\/1906.07315"},{"key":"e_1_2_1_18_1","unstructured":"Anirudh Khatry Joyce Cahoon Jordan Henkel Shaleen Deep Venkatesh Emani Avrilia Floratou Sumit Gulwani Vu Le Mohammad Raza Sherry Shi Mukul Singh and Ashish Tiwari. 2023. From Words to Code: Harnessing Data for Program Synthesis from Natural Language. arXiv:2305.01598 [cs.DB]"},{"key":"e_1_2_1_19_1","doi-asserted-by":"crossref","unstructured":"Jiale Lao Yibo Wang Yufei Li Jianping Wang Yunjia Zhang Zhiyuan Cheng Wanghu Chen Mingjie Tang and Jianguo Wang. 2023. GPTuner: A Manual-Reading Database Tuning System via GPT-Guided Bayesian Optimization. arXiv:2311.03157 [cs.DB]","DOI":"10.1145\/3626246.3654739"},{"key":"e_1_2_1_20_1","unstructured":"Jinyang Li Binyuan Hui Ge Qu Binhua Li Jiaxi Yang Bowen Li Bailin Wang Bowen Qin Rongyu Cao Ruiying Geng et al. 2023. Can llm already serve as a database interface? a big bench for large-scale database grounded text-to-sqls. arXiv preprint arXiv:2305.03111 (2023)."},{"key":"e_1_2_1_21_1","volume-title":"Prefix-tuning: Optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190","author":"Li Xiang Lisa","year":"2021","unstructured":"Xiang Lisa Li and Percy Liang. 2021. Prefix-tuning: Optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190 (2021)."},{"key":"e_1_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"e_1_2_1_23_1","volume-title":"Memory Augmented Policy Optimization for Program Synthesis with Generalization. CoRR abs\/1807.02322","author":"Liang Chen","year":"2018","unstructured":"Chen Liang, Mohammad Norouzi, Jonathan Berant, Quoc V. Le, and Ni Lao. 2018. Memory Augmented Policy Optimization for Program Synthesis with Generalization. CoRR abs\/1807.02322 (2018). arXiv:1807.02322 http:\/\/arxiv.org\/abs\/1807.02322"},{"key":"e_1_2_1_24_1","volume-title":"Rouge: A package for automatic evaluation of summaries. In Text summarization branches out. 74--81.","author":"Lin Chin-Yew","year":"2004","unstructured":"Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out. 74--81."},{"key":"e_1_2_1_25_1","volume-title":"TAPEX: Table Pre-training via Learning a Neural SQL Executor. arXiv:2107.07653 [cs.CL]","author":"Liu Qian","year":"2022","unstructured":"Qian Liu, Bei Chen, Jiaqi Guo, Morteza Ziyadi, Zeqi Lin, Weizhu Chen, and Jian-Guang Lou. 2022. TAPEX: Table Pre-training via Learning a Neural SQL Executor. arXiv:2107.07653 [cs.CL]"},{"key":"e_1_2_1_26_1","volume-title":"Proceedings of the 9th Python in Science Conference","volume":"445","author":"Wes","unstructured":"Wes McKinney et al. 2010. Data structures for statistical computing in python. In Proceedings of the 9th Python in Science Conference, Vol. 445. Austin, TX, 51--56."},{"key":"e_1_2_1_27_1","volume-title":"Power BI 2024","author":"Microsoft","year":"2023","unstructured":"Microsoft Power BI 2024. Microsoft Power BI. Retrieved Jan 27, 2023 from https:\/\/msit.powerbi.com\/home?experience=power-bi"},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00446"},{"key":"e_1_2_1_29_1","volume-title":"Sida I. Wang, and Xi Victoria Lin.","author":"Ni Ansong","year":"2023","unstructured":"Ansong Ni, Srini Iyer, Dragomir Radev, Ves Stoyanov, Wen tau Yih, Sida I. Wang, and Xi Victoria Lin. 2023. LEVER: Learning to Verify Language-to-Code Generation with Execution. arXiv:2302.08468 [cs.LG]"},{"key":"e_1_2_1_31_1","volume-title":"Retrieved","author":"AI","year":"2023","unstructured":"OpenAI models 2023. OpenAI models. Retrieved Sep 17, 2023 from https:\/\/platform.openai.com\/docs\/models"},{"key":"e_1_2_1_32_1","doi-asserted-by":"crossref","unstructured":"Panupong Pasupat and Percy Liang. 2015. Compositional Semantic Parsing on Semi-Structured Tables. arXiv:1508.00305 [cs.CL]","DOI":"10.3115\/v1\/P15-1142"},{"key":"e_1_2_1_33_1","volume-title":"True few-shot learning with language models. Advances in neural information processing systems 34","author":"Perez Ethan","year":"2021","unstructured":"Ethan Perez, Douwe Kiela, and Kyunghyun Cho. 2021. True few-shot learning with language models. Advances in neural information processing systems 34 (2021), 11054--11070."},{"key":"e_1_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3552490.3552496"},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P15-1085"},{"key":"e_1_2_1_36_1","first-page":"1","article-title":"Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer","volume":"21","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. Journal of Machine Learning Research 21, 140 (2020), 1--67. http:\/\/jmlr.org\/papers\/v21\/20-074.html","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_2_1_37_1","volume-title":"Liu","author":"Raffel Colin","year":"2023","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2023. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. arXiv:1910.10683 [cs.LG]"},{"key":"e_1_2_1_38_1","volume-title":"Christopher Michael Rytting, and David Wingate","author":"Robinson Joshua","year":"2022","unstructured":"Joshua Robinson, Christopher Michael Rytting, and David Wingate. 2022. Leveraging large language models for multiple choice question answering. arXiv preprint arXiv:2210.12353 (2022)."},{"key":"e_1_2_1_39_1","unstructured":"Baptiste Rozi\u00e8re Jonas Gehring Fabian Gloeckle Sten Sootla Itai Gat Xiaoqing Ellen Tan Yossi Adi Jingyu Liu Tal Remez J\u00e9r\u00e9my Rapin Artyom Kozhevnikov Ivan Evtimov Joanna Bitton Manish Bhatt Cristian Canton Ferrer Aaron Grattafiori Wenhan Xiong Alexandre D\u00e9fossez Jade Copet Faisal Azhar Hugo Touvron Louis Martin Nicolas Usunier Thomas Scialom and Gabriel Synnaeve. 2023. Code Llama: Open Foundation Models for Code. arXiv:2308.12950 [cs.CL]"},{"key":"e_1_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.779"},{"key":"e_1_2_1_41_1","volume-title":"Natural language to code translation with execution. arXiv preprint arXiv:2204.11454","author":"Shi Freda","year":"2022","unstructured":"Freda Shi, Daniel Fried, Marjan Ghazvininejad, Luke Zettlemoyer, and Sida I Wang. 2022. Natural language to code translation with execution. arXiv preprint arXiv:2204.11454 (2022)."},{"key":"e_1_2_1_42_1","volume-title":"Weijia Shi, Tianlu Wang, Jiayi Xin, Rui Zhang, Mari Ostendorf, Luke Zettlemoyer, Noah A Smith, et al.","author":"Su Hongjin","year":"2022","unstructured":"Hongjin Su, Jungo Kasai, Chen Henry Wu, Weijia Shi, Tianlu Wang, Jiayi Xin, Rui Zhang, Mari Ostendorf, Luke Zettlemoyer, Noah A Smith, et al. 2022. Selective annotation makes language models better few-shot learners. arXiv preprint arXiv:2209.01975 (2022)."},{"key":"e_1_2_1_43_1","volume-title":"Sequence to sequence learning with neural networks. Advances in neural information processing systems 27","author":"Sutskever Ilya","year":"2014","unstructured":"Ilya Sutskever, Oriol Vinyals, and Quoc V Le. 2014. Sequence to sequence learning with neural networks. Advances in neural information processing systems 27 (2014)."},{"key":"e_1_2_1_44_1","volume-title":"Retrieved","author":"Temperature","year":"2023","unstructured":"Temperature setup of GPT models 2023. Temperature setup of GPT models. Retrieved Sep 28, 2023 from https:\/\/platform.openai.com\/docs\/api-reference\/completions"},{"key":"e_1_2_1_45_1","volume-title":"Retrieved","author":"The","year":"2023","unstructured":"The Codex model of OpenAI 2023. The Codex model of OpenAI. Retrieved Sep 28, 2023 from https:\/\/openai.com\/blog\/openai-codex"},{"key":"e_1_2_1_46_1","volume-title":"Retrieved","author":"The","year":"2023","unstructured":"The prompting cost of OpenAI 2023. The prompting cost of OpenAI. Retrieved Dec 22, 2023 from https:\/\/openai.com\/pricing"},{"key":"e_1_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3514221.3517843"},{"key":"e_1_2_1_48_1","volume-title":"Proceedings of the Conference on Innovative Data Systems Research (CIDR). https:\/\/www.cidrdb.org\/cidr2024\/papers\/p14-urban.pdf","author":"Urban Matthias","year":"2024","unstructured":"Matthias Urban and Carsten Binnig. 2024. CAESURA: Language Models as Multi-Modal Query Planners. In Proceedings of the Conference on Innovative Data Systems Research (CIDR). https:\/\/www.cidrdb.org\/cidr2024\/papers\/p14-urban.pdf"},{"key":"e_1_2_1_49_1","volume-title":"Retrieved","author":"Usage","year":"2023","unstructured":"Usage of GPT-3.5-Turbo and GPT-4 Models 2023. Usage of GPT-3.5-Turbo and GPT-4 Models. Retrieved Sep 17, 2023 from https:\/\/learn.microsoft.com\/en-us\/azure\/ai-services\/openai\/how-to\/chatgpt?pivots=programming-language-chat-completions"},{"key":"e_1_2_1_50_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N. Gomez Lukasz Kaiser and Illia Polosukhin. 2023. Attention Is All You Need. arXiv:1706.03762 [cs.CL]"},{"key":"e_1_2_1_51_1","volume-title":"Chi, Quoc Le, and Denny Zhou","author":"Wei Jason","year":"2023","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed Chi, Quoc Le, and Denny Zhou. 2023. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. arXiv:2201.11903 [cs.CL]"},{"key":"e_1_2_1_52_1","volume-title":"Retrieved","author":"TableQuestions","year":"2023","unstructured":"WikiTableQuestions data set 2023. WikiTableQuestions data set. Retrieved Sep 28, 2023 from https:\/\/github.com\/ppasupat\/WikiTableQuestions"},{"key":"e_1_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.40"},{"key":"e_1_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.628"},{"key":"e_1_2_1_55_1","unstructured":"Shunyu Yao Jeffrey Zhao Dian Yu Nan Du Izhak Shafran Karthik Narasimhan and Yuan Cao. 2023. ReAct: Synergizing Reasoning and Acting in Language Models. arXiv:2210.03629 [cs.CL]"},{"key":"e_1_2_1_56_1","unstructured":"Yunhu Ye Binyuan Hui Min Yang Binhua Li Fei Huang and Yongbin Li. 2023. Large Language Models are Versatile Decomposers: Decompose Evidence and Questions for Table-based Reasoning. arXiv:2301.13808 [cs.CL]"},{"key":"e_1_2_1_57_1","volume-title":"Suyi Li, Eric Xue, Bo Pang, Xi Victoria Lin, Yi Chern Tan, Tianze Shi, Zihan Li, et al.","author":"Yu Tao","year":"2019","unstructured":"Tao Yu, Rui Zhang, He Yang Er, Suyi Li, Eric Xue, Bo Pang, Xi Victoria Lin, Yi Chern Tan, Tianze Shi, Zihan Li, et al. 2019. Cosql: A conversational text-to-sql challenge towards cross-domain natural language interfaces to databases. arXiv preprint arXiv:1909.05378 (2019)."},{"key":"e_1_2_1_58_1","volume-title":"Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-sql task. arXiv preprint arXiv:1809.08887","author":"Yu Tao","year":"2018","unstructured":"Tao Yu, Rui Zhang, Kai Yang, Michihiro Yasunaga, Dongxu Wang, Zifan Li, James Ma, Irene Li, Qingning Yao, Shanelle Roman, et al. 2018. Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-sql task. arXiv preprint arXiv:1809.08887 (2018)."},{"key":"e_1_2_1_59_1","volume-title":"Xi Victoria Lin, Suyi Li, Heyang Er, Irene Li, Bo Pang, Tao Chen, et al.","author":"Yu Tao","year":"2019","unstructured":"Tao Yu, Rui Zhang, Michihiro Yasunaga, Yi Chern Tan, Xi Victoria Lin, Suyi Li, Heyang Er, Irene Li, Bo Pang, Tao Chen, et al. 2019. Sparc: Cross-domain semantic parsing in context. arXiv preprint arXiv:1906.02285 (2019)."},{"key":"e_1_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.126"},{"key":"e_1_2_1_61_1","unstructured":"Minjia Zhang Conglong Li Xiaoxia Wu Zhewei Yao and Yuxiong He. 2022. SaMoE: Parameter Efficient MoE Language Models via Self-Adaptive Expert Combination. (2022)."},{"key":"e_1_2_1_62_1","volume-title":"Differentiable prompt makes pre-trained language models better few-shot learners. arXiv preprint arXiv:2108.13161","author":"Zhang Ningyu","year":"2021","unstructured":"Ningyu Zhang, Luoqiu Li, Xiang Chen, Shumin Deng, Zhen Bi, Chuanqi Tan, Fei Huang, and Huajun Chen. 2021. Differentiable prompt makes pre-trained language models better few-shot learners. arXiv preprint arXiv:2108.13161 (2021)."},{"key":"e_1_2_1_63_1","volume-title":"Patel","author":"Zhang Yunjia","year":"2023","unstructured":"Yunjia Zhang, Avrilia Floratou, Joyce Cahoon, Subru Krishnan, Andreas C. M\u00fcller, Dalitso Banda, Fotis Psallidas, and Jignesh M. Patel. 2023. Schema Matching using Pre-Trained Language Models. In ICDE. IEEE. https:\/\/www.microsoft.com\/en-us\/research\/publication\/schema-matching-using-pre-trained-language-models\/"},{"key":"e_1_2_1_64_1","volume-title":"Seq2sql: Generating structured queries from natural language using reinforcement learning. arXiv preprint arXiv:1709.00103","author":"Zhong Victor","year":"2017","unstructured":"Victor Zhong, Caiming Xiong, and Richard Socher. 2017. Seq2sql: Generating structured queries from natural language using reinforcement learning. arXiv preprint arXiv:1709.00103 (2017)."},{"key":"e_1_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.539"},{"key":"e_1_2_1_66_1","doi-asserted-by":"crossref","unstructured":"Fan Zhou Mengkang Hu Haoyu Dong Zhoujun Cheng Shi Han and Dongmei Zhang. 2022. TaCube: Pre-computing Data Cubes for Answering Numerical-Reasoning Questions over Tabular Data. arXiv:2205.12682 [cs.IR]","DOI":"10.18653\/v1\/2022.emnlp-main.145"},{"key":"e_1_2_1_67_1","unstructured":"Wenhao Zhu Hongyi Liu Qingxiu Dong Jingjing Xu Shujian Huang Lingpeng Kong Jiajun Chen and Lei Li. 2023. Multilingual Machine Translation with Large Language Models: Empirical Results and Analysis. arXiv:2304.04675 [cs.CL]"}],"container-title":["Proceedings of the VLDB Endowment"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.14778\/3659437.3659452","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,31]],"date-time":"2024-05-31T16:32:43Z","timestamp":1717173163000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.14778\/3659437.3659452"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4]]},"references-count":66,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["10.14778\/3659437.3659452"],"URL":"https:\/\/doi.org\/10.14778\/3659437.3659452","relation":{},"ISSN":["2150-8097"],"issn-type":[{"value":"2150-8097","type":"print"}],"subject":[],"published":{"date-parts":[[2024,4]]},"assertion":[{"value":"2024-05-31","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}