{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T03:14:48Z","timestamp":1773976488116,"version":"3.50.1"},"reference-count":80,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2025,6,6]],"date-time":"2025-06-06T00:00:00Z","timestamp":1749168000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,6,6]],"date-time":"2025-06-06T00:00:00Z","timestamp":1749168000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001773","name":"University of New South Wales","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100001773","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["World Wide Web"],"published-print":{"date-parts":[[2025,7]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>Hybrid tabular-textual question answering (HTQA) involves integrating multiple data sources, traditionally managed through LSTM-based step-by-step reasoning. However, such sequential approaches are prone to exposure bias and cumulative errors, limiting their effectiveness. This paper first introduces an innovative parallel program generation method, ConcurGen, aiming to transform this paradigm by simultaneously formulating comprehensive program constructs that seamlessly blend operations and values. This approach not only rectifies the inherent pitfalls of sequential methodologies but also infuses efficiency into the process. Through our further research, we found that some HTQA scenarios extend beyond traditional question-answering, often involving open-ended questions that demand dynamic, context-aware response generation. Therefore, we introduce a second framework that leverages large language models (LLMs) to effectively answer both traditional and open-ended questions. Our method demonstrates substantial improvements over existing models such as FinQANet and MT2Net on benchmarks including ConvFinQA and MultiHiertt, achieving new state-of-the-art performance across multiple evaluation metrics. In addition to its accuracy, it delivers a nearly 21x speedup in program generation, significantly enhancing inference efficiency. Unlike traditional models, our system maintains robust performance as the complexity of numerical reasoning increases, highlighting its adaptability in challenging scenarios. Furthermore, supplementary experiments on the LLM-based framework show that it provides enriched answer justifications while achieving similar performance to ConcurGen on standard benchmarks.<\/jats:p>","DOI":"10.1007\/s11280-025-01351-1","type":"journal-article","created":{"date-parts":[[2025,6,5]],"date-time":"2025-06-05T22:14:05Z","timestamp":1749161645000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Tabular-textual question answering: From parallel program generation to large language models"],"prefix":"10.1007","volume":"28","author":[{"given":"Xushuo","family":"Tang","sequence":"first","affiliation":[]},{"given":"Liuyi","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Wenke","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Zhengyi","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Mingchen","family":"Ju","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Shu","sequence":"additional","affiliation":[]},{"given":"Zihan","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Yifu","family":"Tang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,6]]},"reference":[{"key":"1351_CR1","doi-asserted-by":"crossref","unstructured":"Zhu, G., Lin, X., Zhu, K., Zhang, W., Yu, J.X.: Treespan: efficiently computing similarity all-matching. In: Proceedings of the 2012 ACM SIGMOD international conference on management of data, pp 529\u2013540 (2012)","DOI":"10.1145\/2213836.2213896"},{"key":"1351_CR2","doi-asserted-by":"crossref","unstructured":"Qin, L., Peng, Y., Zhang, Y., Lin, X., Zhang, W., Zhou, J.: Towards bridging theory and practice: hop-constrained st simple path enumeration. In: International conference on very large data bases (2019)","DOI":"10.14778\/3372716.3372720"},{"key":"1351_CR3","doi-asserted-by":"publisher","unstructured":"Sun, Q., Lin, X., Zhang, Y., Zhang, W., Chen, C.: Towards higher-order topological consistency for unsupervised network alignment. In: 2023 IEEE 39th international conference on data engineering (ICDE), pp 177\u2013190 (2023). https:\/\/doi.org\/10.1109\/ICDE55515.2023.00021","DOI":"10.1109\/ICDE55515.2023.00021"},{"key":"1351_CR4","doi-asserted-by":"crossref","unstructured":"Wu, Y., Xu, Y., Zhang, W., Xu, X., Zhang, Y.: Query2gmm: Learning representation with gaussian mixture model for reasoning over knowledge graphs. In: Proceedings of the ACM Web Conference 2024, WWW \u201924, Association for Computing Machinery, New York, NY, USA, pp 2149\u20132158 (2024)","DOI":"10.1145\/3589334.3645569"},{"key":"1351_CR5","doi-asserted-by":"publisher","first-page":"703","DOI":"10.1007\/s00778-012-0265-y","volume":"21","author":"MA Cheema","year":"2012","unstructured":"Cheema, M.A., Zhang, W., Lin, X., Zhang, Y.: Efficiently processing snapshot and continuous reverse k nearest neighbors queries. VLDB J. 21, 703\u2013728 (2012)","journal-title":"VLDB J."},{"key":"1351_CR6","doi-asserted-by":"crossref","unstructured":"Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: SQuAD: 100,000+ questions for machine comprehension of text. In: Proceedings of the 2016 conference on empirical methods in natural language processing, pp 2383\u20132392 (2016)","DOI":"10.18653\/v1\/D16-1264"},{"key":"1351_CR7","doi-asserted-by":"crossref","unstructured":"Yang, Z., Qi, P., Zhang, S., Bengio, Y., Cohen, W., Salakhutdinov, R., Manning, C.D.: HotpotQA: A dataset for diverse, explainable multi-hop question answering. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 2369\u20132380 (2018)","DOI":"10.18653\/v1\/D18-1259"},{"key":"1351_CR8","unstructured":"Dua, D., Wang, Y., Dasigi, P., Stanovsky, G., Singh, S., Gardner, M.: DROP: A reading comprehension benchmark requiring discrete reasoning over paragraphs. In: Proceedings of the 2019 conference of the north american chapter of the association for computational linguistics: Human language technologies, pp 2368\u20132378 (2019)"},{"key":"1351_CR9","doi-asserted-by":"crossref","unstructured":"Herzig, J., Nowak, P.K., M\u00fcller, T., Piccinno, F., Eisenschlos, J.: TaPas: Weakly supervised table parsing via pre-training. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 4320\u20134333 (2020)","DOI":"10.18653\/v1\/2020.acl-main.398"},{"key":"1351_CR10","doi-asserted-by":"crossref","unstructured":"Yang, J., Gupta, A., Upadhyay, S., He, L., Goel, R., Paul, S.: Tableformer: Robust transformer modeling for table-text encoding. In: Proceedings of the 60th annual meeting of the association for computational linguistics, ACL 2022, pp 528\u2013537 (2022)","DOI":"10.18653\/v1\/2022.acl-long.40"},{"key":"1351_CR11","doi-asserted-by":"crossref","unstructured":"Chen, W., Zha, H., Chen, Z., Xiong, W., Wang, H., Wang, W.Y.: HybridQA: A dataset of multi-hop question answering over tabular and textual data. In: Findings of the association for computational linguistics: EMNLP 2020, pp 1026\u20131036 (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.91"},{"key":"1351_CR12","doi-asserted-by":"crossref","unstructured":"Zhu, F., Lei, W., Huang, Y., Wang, C., Zhang, S., Lv, J., Feng, F., Chua, T.-S.: TAT-QA: A question answering benchmark on a hybrid of tabular and textual content in finance. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing, pp 3277\u20133287 (2021)","DOI":"10.18653\/v1\/2021.acl-long.254"},{"key":"1351_CR13","doi-asserted-by":"crossref","unstructured":"Chen, Z., Chen, W., Smiley, C., Shah, S., Borova, I., Langdon, D., Moussa, R., Beane, M., Huang, T., Routledge, B.R., Wang, W.Y.: Finqa: A dataset of numerical reasoning over financial data. In: Proceedings of the 2021 conference on empirical methods in natural language processing, EMNLP 2021, pp 3697\u20133711 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.300"},{"key":"1351_CR14","doi-asserted-by":"crossref","unstructured":"Li, M., Feng, F., Zhang, H., He, X., Zhu, F., Chua, T.: Learning to imagine: Integrating counterfactual thinking in neural discrete reasoning. In: Proceedings of the 60th annual meeting of the association for computational linguistics, ACL 2022, pp 57\u201369 (2022)","DOI":"10.18653\/v1\/2022.acl-long.5"},{"key":"1351_CR15","doi-asserted-by":"crossref","unstructured":"Chen, Z., Li, S., Smiley, C., Ma, Z., Shah, S., Wang, W.Y.: Convfinqa: Exploring the chain of numerical reasoning in conversational finance question answering. In: Proceedings of the 2022 conference on empirical methods in natural language processing, EMNLP 2022, pp 6279\u20136292 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.421"},{"key":"1351_CR16","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Li, Y., Li, C., Zhang, R.: Multihiertt: Numerical reasoning over multi hierarchical tabular and textual data. In: Proceedings of the 60th annual meeting of the association for computational linguistics, ACL 2022, pp 6588\u20136600 (2022)","DOI":"10.18653\/v1\/2022.acl-long.454"},{"key":"1351_CR17","unstructured":"An, Y.L.: Roberta: A robustly optimized BERT pretraining approach. ArXiv preprint arXiv:1907.11692 (2019)"},{"key":"1351_CR18","doi-asserted-by":"crossref","unstructured":"Chen, L., Han, B., Wang, X., Zhao, J., Yang, W., Yang, Z.: Machine learning methods in weather and climate applications: A survey. Appl. Sci. 13(21) (2023)","DOI":"10.3390\/app132112019"},{"key":"1351_CR19","doi-asserted-by":"crossref","unstructured":"He, Z., Wang, X., Wang, R., Shi, S., Tu, Z.: Bridging the data gap between training and inference for unsupervised neural machine translation. In: Proceedings of the 60th annual meeting of the association for computational linguistics, ACL 2022, pp 6611\u20136623 (2022)","DOI":"10.18653\/v1\/2022.acl-long.456"},{"key":"1351_CR20","doi-asserted-by":"crossref","unstructured":"Zhang, S., Huang, H., Liu, J., Li, H.: Spelling error correction with soft-masked BERT. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 882\u2013890 (2020)","DOI":"10.18653\/v1\/2020.acl-main.82"},{"key":"1351_CR21","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al.: Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1351_CR22","doi-asserted-by":"crossref","unstructured":"Wu, J., Tang, X., Yang, Z., Hao, K., Lai, L., Liu, Y.: An experimental evaluation of & nbsp;llm on & nbsp;image classification, Springer, Berlin, Heidelberg, pp 506\u2013518 (2024)","DOI":"10.1007\/978-981-96-1242-0_37"},{"issue":"9","key":"1351_CR23","doi-asserted-by":"publisher","first-page":"2227","DOI":"10.14778\/3665844.3665853","volume":"17","author":"J Wang","year":"2024","unstructured":"Wang, J., Wang, K., Lin, X., Zhang, W., Zhang, Y.: Efficient unsupervised community search with pre-trained graph transformer. Proc. VLDB Endow. 17(9), 2227\u20132240 (2024)","journal-title":"Proc. VLDB Endow."},{"key":"1351_CR24","doi-asserted-by":"crossref","unstructured":"Yang, W., Yang, Z., Chen, L., Yan, R., Yang, Z., Zhang, L., Tang, Y.: Parallel program generation for hybrid tabular-textual question answering. In: Asia-pacific Web (APWeb) and Web-Age information management (WAIM) joint international conference on Web and big data, Springer, pp 121\u2013137 (2024)","DOI":"10.1007\/978-981-97-7232-2_9"},{"key":"1351_CR25","unstructured":"Hermann, K.M., Kocisky, T., Grefenstette, E., Espeholt, L., Kay, W., Suleyman, M., Blunsom, P.: Teaching machines to read and comprehend. In: NIPS (2015)"},{"key":"1351_CR26","doi-asserted-by":"crossref","unstructured":"Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: SQuAD: 100,000+ questions for machine comprehension of text. In: Proceedings of the 2016 conference on empirical methods in natural language processing, pp 2383\u20132392 (2016)","DOI":"10.18653\/v1\/D16-1264"},{"key":"1351_CR27","doi-asserted-by":"crossref","unstructured":"Berant, J., Chou, A., Frostig, R., Liang, P.: Semantic parsing on Freebase from question-answer pairs. In: Proceedings of the 2013 conference on empirical methods in natural language processing, pp 1533\u20131544 (2013)","DOI":"10.18653\/v1\/D13-1160"},{"key":"1351_CR28","doi-asserted-by":"crossref","unstructured":"Pasupat, P., Liang, P.: Compositional semantic parsing on semi-structured tables. In: Proceedings of the 53rd annual meeting of the association for computational linguistics and the 7th international joint conference on natural language processing, pp 1470\u20131480 (2015)","DOI":"10.3115\/v1\/P15-1142"},{"key":"1351_CR29","doi-asserted-by":"crossref","unstructured":"Yu, T., Zhang, R., Yang, K., Yasunaga, M., Wang, D., Li, Z., Ma, J., Li, I., Yao, Q., Roman, S., Zhang, Z., Radev, D.: Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-SQL task. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 3911\u20133921 (2018)","DOI":"10.18653\/v1\/D18-1425"},{"key":"1351_CR30","doi-asserted-by":"publisher","unstructured":"Wu, Y., Xu, Y., Lin, X., Zhang, W.: A holistic approach for answering logical queries on knowledge graphs. In: 2023 IEEE 39th international conference on data engineering (ICDE), pp 2345\u20132357 (2023). https:\/\/doi.org\/10.1109\/ICDE55515.2023.00181","DOI":"10.1109\/ICDE55515.2023.00181"},{"key":"1351_CR31","doi-asserted-by":"publisher","unstructured":"Ukey, N., Yang, Z., Li, B., Zhang, G., Hu, Y., Zhang, W.: Survey on exact knn queries over high-dimensional data space. Sensors 23(2) (2023). https:\/\/doi.org\/10.3390\/s23020629","DOI":"10.3390\/s23020629"},{"key":"1351_CR32","unstructured":"Dua, D., Wang, Y., Dasigi, P., Stanovsky, G., Singh, S., Gardner, M.: DROP: A reading comprehension benchmark requiring discrete reasoning over paragraphs. In: Proceedings of the 2019 conference of the north american chapter of the association for computational linguistics: Human language technologies, pp 2368\u20132378 (2019)"},{"key":"1351_CR33","doi-asserted-by":"crossref","unstructured":"Pal, K.K., Baral, C.: Investigating numeracy learning ability of a text-to-text transfer model. In: Findings of the association for computational linguistics: EMNLP, pp 3095\u20133101 (2021)","DOI":"10.18653\/v1\/2021.findings-emnlp.265"},{"key":"1351_CR34","doi-asserted-by":"crossref","unstructured":"Ran, Q., Lin, Y., Li, P., Zhou, J., Liu, Z.: NumNet: Machine reading comprehension with numerical reasoning. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pp 2474\u20132484 (2019)","DOI":"10.18653\/v1\/D19-1251"},{"key":"1351_CR35","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Wang, L., Yu, S., Wang, S., Wang, Y., Jiang, J., Lim, E.: NOAHQA: numerical reasoning with interpretable graph question answering dataset. In: Findings of the association for computational linguistics: EMNLP2021, pp 4147\u20134161 (2021)","DOI":"10.18653\/v1\/2021.findings-emnlp.350"},{"issue":"15","key":"1351_CR36","first-page":"13297","volume":"35","author":"X Li","year":"2021","unstructured":"Li, X., Sun, Y., Cheng, G.: Tsqa: Tabular scenario based question answering. Proc. AAAI Conf. Artif. Intell. 35(15), 13297\u201313305 (2021)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"1351_CR37","doi-asserted-by":"crossref","unstructured":"Deng, Y., Lei, W., Zhang, W., Lam, W., Chua, T.: PACIFIC: towards proactive conversational question answering over tabular and textual data in finance. In: Proceedings of the 2022 conference on empirical methods in natural language processing, EMNLP 2022, pp 6970\u20136984 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.469"},{"key":"1351_CR38","doi-asserted-by":"crossref","unstructured":"Geva, M., Gupta, A., Berant, J.: Injecting numerical reasoning skills into language models. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 946\u2013958 (2020)","DOI":"10.18653\/v1\/2020.acl-main.89"},{"key":"1351_CR39","doi-asserted-by":"crossref","unstructured":"Berg-Kirkpatrick, T., Spokoyny, D.: An empirical investigation of contextualized number prediction. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), pp 4754\u20134764 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.385"},{"key":"1351_CR40","doi-asserted-by":"crossref","unstructured":"Pi, X., Liu, Q., Chen, B., Ziyadi, M., Lin, Z., Fu, Q., Gao, Y., Lou, J., Chen, W.: Reasoning like program executors. In: Proceedings of the 2022 conference on empirical methods in natural language processing, EMNLP 2022, pp 761\u2013779 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.48"},{"key":"1351_CR41","unstructured":"Zhuang, L., Wayne, L., Ya, S., Jun, Z.: A robustly optimized BERT pre-training approach with post-training. In: Li, S., Sun, M., Liu, Y., Wu, H., Liu, K., Che, W., He, S., Rao, G. (eds.) Proceedings of the 20th chinese national conference on computational linguistics, pp 1218\u20131227 (2021)"},{"key":"1351_CR42","unstructured":"Guu, K., Lee, K., Tung, Z., Pasupat, P., Chang, M.: Retrieval augmented language model pre-training. In: International conference on machine learning, PMLR, pp 3929\u20133938 (2020)"},{"key":"1351_CR43","doi-asserted-by":"crossref","unstructured":"Liang, Z., Zhang, J., Wang, L., Qin, W., Lan, Y., Shao, J., Zhang, X.: MWP-BERT: Numeracy-augmented pre-training for math word problem solving. In: Findings of the association for computational linguistics: NAACL 2022, pp 997\u20131009 (2022)","DOI":"10.18653\/v1\/2022.findings-naacl.74"},{"key":"1351_CR44","doi-asserted-by":"crossref","unstructured":"Mishra, S., Finlayson, M., Lu, P., Tang, L., Welleck, S., Baral, C., Rajpurohit, T., Tafjord, O., Sabharwal, A., Clark, P., Kalyan, A.: LILA: A unified benchmark for mathematical reasoning. In: Goldberg, Y., Kozareva, Z., Zhang, Y. (eds.) Proceedings of the 2022 conference on empirical methods in natural language processing, Association for Computational Linguistics, Abu Dhabi, United Arab Emirates, pp 5807\u20135832 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.392"},{"issue":"4","key":"1351_CR45","first-page":"39","volume":"41","author":"P Clark","year":"2020","unstructured":"Clark, P., Etzioni, O., Khashabi, D., Khot, T., Mishra, B.D., Richardson, K., Sabharwal, A., Schoenick, C., Tafjord, O., Tandon, N., Bhakthavatsalam, S., Groeneveld, D., Guerquin, M., Schmitz, M.: From f to a on the new york regents science exams \u2014 an overview of the aristo project. AI Mag. 41(4), 39\u201353 (2020)","journal-title":"AI Mag."},{"key":"1351_CR46","unstructured":"Welleck, S., Liu, J., Lu, X., Hajishirzi, H., Choi, Y.: Naturalprover: grounded mathematical proof generation with language models. In: Proceedings of the 36th international conference on neural information processing systems. NIPS \u201922, Curran Associates Inc., Red Hook, NY, USA (2022)"},{"key":"1351_CR47","first-page":"140","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., Liu, P.J.: Exploring the limits of transfer learning with a unified text-to-text transformer 21, 140\u2013114067 (2020)","journal-title":"Exploring the limits of transfer learning with a unified text-to-text transformer"},{"key":"1351_CR48","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. In: 7th International conference on learning representations, ICLR 2019 (2019)"},{"key":"1351_CR49","doi-asserted-by":"crossref","unstructured":"Yang, P., Wang, H., Zhang, Y., Qin, L., Zhang, W., Lin, X.: T3s: Effective representation learning for trajectory similarity computation. In: 2021 IEEE 37th international conference on data engineering (ICDE), pp 2183\u20132188 (2021)","DOI":"10.1109\/ICDE51399.2021.00221"},{"key":"1351_CR50","unstructured":"Baldi, P., Sadowski, P.J.: Understanding dropout. In: Advances in Neural Information Processing Systems 26: 27th Annual conference on neural information processing systems 2013, pp 2814\u20132822 (2013)"},{"key":"1351_CR51","doi-asserted-by":"crossref","unstructured":"Chen, W., Zha, H., Chen, Z., Xiong, W., Wang, H., Wang, W.Y.: HybridQA: A dataset of multi-hop question answering over tabular and textual data. In: Findings of the association for computational linguistics: EMNLP 2020, pp 1026\u20131036 (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.91"},{"issue":"2","key":"1351_CR52","doi-asserted-by":"publisher","first-page":"205395172110477","DOI":"10.1177\/20539517211047734","volume":"8","author":"D Luitse","year":"2021","unstructured":"Luitse, D., Denkena, W.: The great transformer: Examining the role of large language models in the political economy of ai. Big Data Soc. 8(2), 20539517211047736 (2021)","journal-title":"Big Data Soc."},{"key":"1351_CR53","unstructured":"Naveed, H., Khan, A.U., Qiu, S., Saqib, M., Anwar, S., Usman, M., Akhtar, N., Barnes, N., Mian, A.: A comprehensive overview of large language models. arXiv preprint arXiv:2307.06435 (2023)"},{"key":"1351_CR54","volume-title":"Attention is all you need","author":"A Vaswani","year":"2017","unstructured":"Vaswani, A.: Attention is all you need. Adv. Neural Inf. Process, Syst (2017)"},{"key":"1351_CR55","doi-asserted-by":"publisher","DOI":"10.1016\/j.physd.2019.132306","volume":"404","author":"A Sherstinsky","year":"2020","unstructured":"Sherstinsky, A.: Fundamentals of recurrent neural network (rnn) and long short-term memory (lstm) network. Phys. D Nonlinear Pheno. 404, 132306 (2020)","journal-title":"Phys. D Nonlinear Pheno."},{"issue":"7","key":"1351_CR56","doi-asserted-by":"publisher","first-page":"1235","DOI":"10.1162\/neco_a_01199","volume":"31","author":"Y Yu","year":"2019","unstructured":"Yu, Y., Si, X., Hu, C., Zhang, J.: A review of recurrent neural networks: Lstm cells and network architectures. Neural Comput. 31(7), 1235\u20131270 (2019)","journal-title":"Neural Comput."},{"key":"1351_CR57","volume-title":"A survey of knowledge enhanced pre-trained language models","author":"L Hu","year":"2023","unstructured":"Hu, L., Liu, Z., Zhao, Z., Hou, L., Nie, L., Li, J.: A survey of knowledge enhanced pre-trained language models. IEEE Trans. Knowl, Data Eng (2023)"},{"key":"1351_CR58","doi-asserted-by":"publisher","first-page":"681","DOI":"10.1007\/s11023-020-09548-1","volume":"30","author":"L Floridi","year":"2020","unstructured":"Floridi, L., Chiriatti, M.: Gpt-3: Its nature, scope, limits, and consequences. Minds Mach. 30, 681\u2013694 (2020)","journal-title":"Minds Mach."},{"key":"1351_CR59","unstructured":"Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F.L., Almeida, D., Altenschmidt, J., Altman, S., Anadkat, S., et al.: Gpt-4 technical report. arXiv preprint arXiv:2303.08774 (2023)"},{"key":"1351_CR60","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Han, X., Liu, Z., Jiang, X., Sun, M., Liu, Q.: Ernie: Enhanced language representation with informative entities. arXiv preprint arXiv:1905.07129 (2019)","DOI":"10.18653\/v1\/P19-1139"},{"key":"1351_CR61","doi-asserted-by":"crossref","unstructured":"Sun, Y., Wang, S., Li, Y., Feng, S., Tian, H., Wu, H., Wang, H.: Ernie 2.0: A continual pre-training framework for language understanding. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 8968\u20138975 (2020)","DOI":"10.1609\/aaai.v34i05.6428"},{"key":"1351_CR62","unstructured":"Sun, Y., Wang, S., Feng, S., Ding, S., Pang, C., Shang, J., Liu, J., Chen, X., Zhao, Y., Lu, Y., et al.: Ernie 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation. arXiv preprint arXiv:2107.02137 (2021)"},{"key":"1351_CR63","unstructured":"Sun, Y., Wang, S., Li, Y., Feng, S., Chen, X., Zhang, H., Tian, X., Zhu, D., Tian, H., Wu, H.: Ernie: Enhanced representation through knowledge integration. arXiv preprint arXiv:1904.09223 (2019)"},{"key":"1351_CR64","unstructured":"Bai, J., Bai, S., Chu, Y., Cui, Z., Dang, K., Deng, X., Fan, Y., Ge, W., Han, Y., Huang, F., et al.: Qwen technical report. arXiv preprint arXiv:2309.16609 (2023)"},{"key":"1351_CR65","doi-asserted-by":"crossref","unstructured":"Li, Y., Wang, S., Ding, H., Chen, H.: Large language models in finance: A survey. In: Proceedings of the 4th ACM international conference on AI in finance, pp 374\u2013382 (2023)","DOI":"10.1145\/3604237.3626869"},{"key":"1351_CR66","unstructured":"Nie, Y., Kong, Y., Dong, X., Mulvey, J.M., Poor, H.V., Wen, Q., Zohren, S.: A survey of large language models for financial applications: Progress, prospects and challenges. arXiv preprint arXiv:2406.11903 (2024)"},{"key":"1351_CR67","doi-asserted-by":"crossref","unstructured":"Srivastava, P., Malik, M., Ganu, T.: Assessing llms\u2019 mathematical reasoning in financial document question answering. arXiv preprint arXiv:2402.11194 (2024)","DOI":"10.18653\/v1\/2024.findings-acl.231"},{"key":"1351_CR68","doi-asserted-by":"crossref","unstructured":"Srivastava, P., Malik, M., Gupta, V., Ganu, T., Roth, D.: Evaluating llms\u2019 mathematical reasoning in financial document question answering. In: Findings of the association for computational linguistics ACL 2024, pp 3853\u20133878 (2024)","DOI":"10.18653\/v1\/2024.findings-acl.231"},{"key":"1351_CR69","unstructured":"Ge, Y., Hua, W., Mei, K., Tan, J., Xu, S., Li, Z., Zhang, Y., et al.: Openagi: When llm meets domain experts. Adv. Neural Inf. Process. Syst. 36 (2024)"},{"key":"1351_CR70","unstructured":"Wang, Z., Zhong, W., Wang, Y., Zhu, Q., Mi, F., Wang, B., Shang, L., Jiang, X., Liu, Q.: Data management for large language models: A survey. CoRR (2023)"},{"issue":"4","key":"1351_CR71","doi-asserted-by":"publisher","first-page":"255","DOI":"10.1002\/hcs2.61","volume":"2","author":"R Yang","year":"2023","unstructured":"Yang, R., Tan, T.F., Lu, W., Thirunavukarasu, A.J., Ting, D.S.W., Liu, N.: Large language models in health care: Development, applications, and challenges. Health Care Sci. 2(4), 255\u2013263 (2023)","journal-title":"Health Care Sci."},{"key":"1351_CR72","doi-asserted-by":"crossref","unstructured":"Xiao, C., Xu, S.X., Zhang, K., Wang, Y., Xia, L.: Evaluating reading comprehension exercises generated by llms: A showcase of chatgpt in education applications. In: Proceedings of the 18th workshop on innovative use of NLP for building educational applications (BEA 2023), pp 610\u2013625 (2023)","DOI":"10.18653\/v1\/2023.bea-1.52"},{"key":"1351_CR73","unstructured":"Kolasani, S.: Optimizing natural language processing, large language models (llms) for efficient customer service, and hyper-personalization to enable sustainable growth and revenue. Trans. Latest Trends Artif. Intell. 4(4) (2023)"},{"key":"1351_CR74","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1017\/ATSIP.2019.12","volume":"8","author":"B Wang","year":"2019","unstructured":"Wang, B., Wang, A., Chen, F., Wang, Y., Kuo, C.-C.J.: Evaluating word embedding models: Methods and experimental results. APSIPA Trans. Signal Inf. Process. 8, 19 (2019)","journal-title":"APSIPA Trans. Signal Inf. Process."},{"key":"1351_CR75","unstructured":"Zhang, P., Xiao, S., Liu, Z., Dou, Z., Nie, J.-Y.: Retrieve anything to augment large language models. arXiv preprint arXiv:2310.07554 (2023)"},{"key":"1351_CR76","doi-asserted-by":"crossref","unstructured":"Lee, J., Stevens, N., Han, S.C., Song, M.: A survey of large language models in finance (finllms). arXiv preprint arXiv:2402.02315 (2024)","DOI":"10.1007\/s00521-024-10495-6"},{"key":"1351_CR77","unstructured":"Clarifai: text-embedding-3-large. https:\/\/clarifai.com\/openai\/embed\/models\/text-embedding-3-large (2024)"},{"key":"1351_CR78","first-page":"30233","volume":"35","author":"A Kusupati","year":"2022","unstructured":"Kusupati, A., Bhatt, G., Rege, A., Wallingford, M., Sinha, A., Ramanujan, V., Howard-Snyder, W., Chen, K., Kakade, S., Jain, P., et al.: Matryoshka representation learning. Adv. Neural Inf. Process. Syst. 35, 30233\u201330249 (2022)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1351_CR79","unstructured":"Cohere: Announcing Rerank-v3.5. https:\/\/docs.cohere.com\/changelog\/rerank-v3.5 (2024)"},{"issue":"8","key":"1351_CR80","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al.: Language models are unsupervised multitask learners. OpenAI blog 1(8), 9 (2019)","journal-title":"OpenAI blog"}],"container-title":["World Wide Web"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11280-025-01351-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11280-025-01351-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11280-025-01351-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T18:07:15Z","timestamp":1757182035000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11280-025-01351-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,6]]},"references-count":80,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,7]]}},"alternative-id":["1351"],"URL":"https:\/\/doi.org\/10.1007\/s11280-025-01351-1","relation":{},"ISSN":["1386-145X","1573-1413"],"issn-type":[{"value":"1386-145X","type":"print"},{"value":"1573-1413","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,6,6]]},"assertion":[{"value":"20 February 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 April 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 May 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 June 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Approval"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"42"}}