{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T18:06:38Z","timestamp":1772906798276,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":24,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,17]],"date-time":"2024-10-17T00:00:00Z","timestamp":1729123200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,17]]},"DOI":"10.1145\/3723178.3723244","type":"proceedings-article","created":{"date-parts":[[2025,6,6]],"date-time":"2025-06-06T07:16:47Z","timestamp":1749194207000},"page":"498-505","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Exploring the Effectiveness of Large Language Models in Financial Question Answering"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-4598-1952","authenticated-orcid":false,"given":"Mondol","family":"Mridul Provakar","sequence":"first","affiliation":[{"name":"Rajshahi University of Engineering &amp; Technology, Rajshahi, Bangladesh"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6575-0193","authenticated-orcid":false,"given":"Emrana Kabir","family":"Hashi","sequence":"additional","affiliation":[{"name":"Rajshahi University of Engineering &amp; Technology, Rajshahi, Bangladesh"}]}],"member":"320","published-online":{"date-parts":[[2025,6,6]]},"reference":[{"key":"e_1_3_3_1_2_2","doi-asserted-by":"publisher","unstructured":"Hessa\u00a0Abdulrahman Alawwad Areej Alhothali Usman Naseem Ali Alkhathlan and Amani Jamal. 2024. Enhancing Textbook Question Answering Task with Large Language Models and Retrieval Augmented Generation. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.05128 (2024). 10.48550\/arXiv.2402.05128","DOI":"10.48550\/arXiv.2402.05128"},{"key":"e_1_3_3_1_3_2","doi-asserted-by":"publisher","unstructured":"Yupeng Chang Xu Wang Jindong Wang Yuan Wu Linyi Yang Kaijie Zhu Hao Chen Xiaoyuan Yi Cunxiang Wang Yidong Wang Wei Ye Yue Zhang Yi Chang Philip\u00a0S. Yu Qiang Yang and Xing Xie. 2024. A Survey on Evaluation of Large Language Models. 15 3 (2024). 10.1145\/3641289","DOI":"10.1145\/3641289"},{"key":"e_1_3_3_1_4_2","doi-asserted-by":"publisher","unstructured":"Zhiyu Chen Wenhu Chen Charese Smiley Sameena Shah Iana Borova Dylan Langdon Reema Moussa Matt Beane Ting-Hao Huang Bryan Routledge et\u00a0al. 2021. Finqa: A dataset of numerical reasoning over financial data. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2109.00122 (2021). 10.48550\/arXiv.2109.00122","DOI":"10.48550\/arXiv.2109.00122"},{"key":"e_1_3_3_1_5_2","doi-asserted-by":"publisher","unstructured":"Zhiyu Chen Shiyang Li Charese Smiley Zhiqiang Ma Sameena Shah and William\u00a0Yang Wang. 2022. Convfinqa: Exploring the chain of numerical reasoning in conversational finance question answering. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2210.03849 (2022). 10.48550\/arXiv.2210.03849","DOI":"10.48550\/arXiv.2210.03849"},{"key":"e_1_3_3_1_6_2","unstructured":"Coldstart Coder. [n.d.]. Getting Started with Google\u2019s Gemma LLM using Huggingface Libraries. https:\/\/medium.com\/@coldstart_coder\/getting-started-with-googles-gemma-llm-using-huggingface-libraries-a0d826c552ae. Medium ([n. d.]). Accessed: 12\/04\/2024."},{"key":"e_1_3_3_1_7_2","doi-asserted-by":"publisher","unstructured":"Ning Ding Yujia Qin Guang Yang Fuchao Wei Zonghan Yang Yusheng Su Shengding Hu Yulin Chen Chi-Min Chan Weize Chen et\u00a0al. 2023. Parameter-efficient fine-tuning of large-scale pre-trained language models. Nature Machine Intelligence 5 3 (2023) 220\u2013235. 10.1038\/s42256-023-00626-4","DOI":"10.1038\/s42256-023-00626-4"},{"key":"e_1_3_3_1_8_2","doi-asserted-by":"publisher","unstructured":"John\u00a0W. Goodell Satish Kumar Weng\u00a0Marc Lim and Debidutta Pattnaik. 2021. Artificial intelligence and machine learning in finance: Identifying foundations themes and research clusters from bibliometric analysis. Journal of Behavioral and Experimental Finance 32 (2021) 100577. 10.1016\/j.jbef.2021.100577","DOI":"10.1016\/j.jbef.2021.100577"},{"key":"e_1_3_3_1_9_2","doi-asserted-by":"publisher","unstructured":"Edward\u00a0J Hu Yelong Shen Phillip Wallis Zeyuan Allen-Zhu Yuanzhi Li Shean Wang Lu Wang and Weizhu Chen. 2021. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2106.09685 (2021). 10.48550\/arXiv.2106.09685","DOI":"10.48550\/arXiv.2106.09685"},{"key":"e_1_3_3_1_10_2","doi-asserted-by":"publisher","unstructured":"Pranab Islam Anand Kannappan Douwe Kiela Rebecca Qian Nino Scherrer and Bertie Vidgen. 2023. Financebench: A new benchmark for financial question answering. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2311.11944 (2023). 10.48550\/arXiv.2311.11944","DOI":"10.48550\/arXiv.2311.11944"},{"key":"e_1_3_3_1_11_2","doi-asserted-by":"publisher","unstructured":"Takeshi Kojima Shixiang\u00a0Shane Gu Machel Reid Yutaka Matsuo and Yusuke Iwasawa. 2022. Large language models are zero-shot reasoners. Advances in neural information processing systems 35 (2022) 22199\u201322213. 10.18653\/v1\/2023.findings-eacl.83","DOI":"10.18653\/v1\/2023.findings-eacl.83"},{"key":"e_1_3_3_1_12_2","doi-asserted-by":"publisher","DOI":"10.1145\/3604237.3626869"},{"key":"e_1_3_3_1_13_2","unstructured":"Monodol007. 2024. Finance Question Answering Correctness Evaluation. https:\/\/github.com\/Mondol007\/Finance_Question_Answering\/blob\/main\/Finance_Question_Answering_Correctness_Evaluation.pdf Accessed: 06\/08\/2024."},{"key":"e_1_3_3_1_14_2","doi-asserted-by":"publisher","unstructured":"Zooey Nguyen Anthony Annunziata Vinh Luong Sang Dinh Quynh Le Anh\u00a0Hai Ha Chanh Le Hong\u00a0An Phan Shruti Raghavan and Christopher Nguyen. 2024. Enhancing Q&A with Domain-Specific Fine-Tuning and Iterative Reasoning: A Comparative Study. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2404.11792 (2024). 10.48550\/arXiv.2404.11792","DOI":"10.48550\/arXiv.2404.11792"},{"key":"e_1_3_3_1_15_2","unstructured":"NLPlanet. 2021. Two Minutes NLP: Learn the ROUGE Metric by Examples. https:\/\/medium.com\/nlplanet\/two-minutes-nlp-learn-the-rouge-metric-by-examples-f179cc285499. Medium (2021). Accessed: 12\/04\/2024."},{"key":"e_1_3_3_1_16_2","unstructured":"Patronus AI. 2022. FinanceBench. https:\/\/huggingface.co\/datasets\/PatronusAI\/financebench. Accessed: 12\/04\/2024."},{"key":"e_1_3_3_1_17_2","unstructured":"Replicate. Year. How to Prompt: Llama. https:\/\/replicate.com\/blog\/how-to-prompt-llama. Accessed: 12\/04\/2024."},{"key":"e_1_3_3_1_18_2","unstructured":"Seaplane. 2024. Evaluating and Measuring LLM Performance: A Quantitative Approach. https:\/\/www.seaplane.io\/blog\/evaluating-and-measure-llm-performance-a-quantitative-approach. Accessed: 12\/04\/2024."},{"key":"e_1_3_3_1_19_2","doi-asserted-by":"publisher","unstructured":"Shijie Wu Ozan Irsoy Steven Lu Vadim Dabravolski Mark Dredze Sebastian Gehrmann Prabhanjan Kambadur David Rosenberg and Gideon Mann. 2023. Bloomberggpt: A large language model for finance. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.17564 (2023). 10.48550\/arXiv.2303.17564","DOI":"10.48550\/arXiv.2303.17564"},{"key":"e_1_3_3_1_20_2","doi-asserted-by":"publisher","unstructured":"Hongyang Yang Xiao-Yang Liu and Christina\u00a0Dan Wang. 2023. Fingpt: Open-source financial large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2306.06031 (2023). 10.48550\/arXiv.2306.06031","DOI":"10.48550\/arXiv.2306.06031"},{"key":"e_1_3_3_1_21_2","doi-asserted-by":"publisher","unstructured":"Xi Ye and Greg Durrett. 2022. The unreliability of explanations in few-shot prompting for textual reasoning. Advances in neural information processing systems 35 (2022) 30378\u201330392. 10.48550\/arXiv.2205.03401","DOI":"10.48550\/arXiv.2205.03401"},{"key":"e_1_3_3_1_22_2","doi-asserted-by":"publisher","unstructured":"Hangwen Zhang Qingyi Si Peng Fu Zheng Lin and Weiping Wang. 2024. Are Large Language Models Table-based Fact-Checkers? arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.02549 (2024). 10.48550\/arXiv.2402.02549","DOI":"10.48550\/arXiv.2402.02549"},{"key":"e_1_3_3_1_23_2","doi-asserted-by":"publisher","unstructured":"Liang Zhang Katherine Jijo Spurthi Setty Eden Chung Fatima Javid Natan Vidra and Tommy Clifford. 2024. Enhancing Large Language Model Performance To Answer Questions and Extract Information More Accurately. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.01722 (2024). 10.48550\/arXiv.2402.01722","DOI":"10.48550\/arXiv.2402.01722"},{"key":"e_1_3_3_1_24_2","doi-asserted-by":"publisher","unstructured":"Zheng Zhang Chen Zheng Da Tang Ke Sun Yukun Ma Yingtong Bu Xun Zhou and Liang Zhao. 2023. Balancing specialized and general skills in llms: The impact of modern tuning and data strategy. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2310.04945 (2023). 10.48550\/arXiv.2310.04945","DOI":"10.48550\/arXiv.2310.04945"},{"key":"e_1_3_3_1_25_2","doi-asserted-by":"publisher","unstructured":"Huaqin Zhao Zhengliang Liu Zihao Wu Yiwei Li Tianze Yang Peng Shu Shaochen Xu and et al.2024. Revolutionizing finance with LLMs: An overview of applications and insights. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2401.11641 (2024). 10.48550\/arXiv.2401.11641","DOI":"10.48550\/arXiv.2401.11641"}],"event":{"name":"ICCA 2024: 3rd International Conference on Computing Advancements","location":"Dhaka Bangladesh","acronym":"ICCA 2024"},"container-title":["Proceedings of the 3rd International Conference on Computing Advancements"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3723178.3723244","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3723178.3723244","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:56:47Z","timestamp":1750298207000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3723178.3723244"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,17]]},"references-count":24,"alternative-id":["10.1145\/3723178.3723244","10.1145\/3723178"],"URL":"https:\/\/doi.org\/10.1145\/3723178.3723244","relation":{},"subject":[],"published":{"date-parts":[[2024,10,17]]},"assertion":[{"value":"2025-06-06","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}