{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T00:55:07Z","timestamp":1773968107947,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":44,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,11,14]],"date-time":"2024-11-14T00:00:00Z","timestamp":1731542400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,11,14]]},"DOI":"10.1145\/3677052.3698675","type":"proceedings-article","created":{"date-parts":[[2024,11,14]],"date-time":"2024-11-14T06:38:06Z","timestamp":1731566286000},"page":"626-633","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Is Small Really Beautiful for Central Bank Communication? Evaluating Language Models for Finance: Llama-3-70B, GPT-4, FinBERT-FOMC, FinBERT, and VADER"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1850-471X","authenticated-orcid":false,"given":"Wonseong","family":"Kim","sequence":"first","affiliation":[{"name":"Institute of Economics and Statistics, Korea University, KR"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9473-5029","authenticated-orcid":false,"given":"Jan","family":"Sp\u00f6rer","sequence":"additional","affiliation":[{"name":"Institute of Computer Science, Chair of Natural Language Processing, University of St. Gallen, CH"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9892-1750","authenticated-orcid":false,"given":"Choong Lyol","family":"Lee","sequence":"additional","affiliation":[{"name":"Institute of Economics and Statistics, Korea University, KR"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6195-9034","authenticated-orcid":false,"given":"Siegfried","family":"Handschuh","sequence":"additional","affiliation":[{"name":"Institute of Computer Science, Chair of Natural Language Processing, University of St.Gallen, CH"}]}],"member":"320","published-online":{"date-parts":[[2024,11,14]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Josh Achiam Steven Adler Sandhini Agarwal Lama Ahmad Ilge Akkaya Florencia\u00a0Leoni Aleman and Diogo Almeida. 2023. GPT-4 Technical Report."},{"key":"e_1_3_2_1_2_1","first-page":"49","article-title":"A New Measure of Central Bank Transparency and Implications for the Effectiveness of Monetary Policy","volume":"19","author":"Acosta Miguel","year":"2023","unstructured":"Miguel Acosta. 2023. A New Measure of Central Bank Transparency and Implications for the Effectiveness of Monetary Policy. International Journal of Central Banking 19, 3 (2023), 49\u201397.","journal-title":"International Journal of Central Banking"},{"key":"e_1_3_2_1_3_1","volume-title":"GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints.","author":"Ainslie Joshua","year":"2023","unstructured":"Joshua Ainslie, James Lee-Thorp, Michiel de Jong, Yury Zemlyanskiy, Federico Lebr\u00f3n, and Sumit Sanghai. 2023. GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints."},{"key":"e_1_3_2_1_4_1","volume-title":"FinBERT: Financial Sentiment Analysis with Pre-trained Language Models. arXiv","author":"Araci Dogu","year":"2019","unstructured":"Dogu Araci. 2019. FinBERT: Financial Sentiment Analysis with Pre-trained Language Models. arXiv (2019)."},{"key":"e_1_3_2_1_5_1","volume-title":"The Information Content of FOMC Minutes. SSRN","author":"Boukus Ellyn","year":"2006","unstructured":"Ellyn Boukus and Joshua\u00a0V. Rosenberg. 2006. The Information Content of FOMC Minutes. SSRN (2006)."},{"key":"e_1_3_2_1_6_1","volume-title":"FinBERT-FOMC: Fine-Tuned FinBERT Model with Sentiment Focus Method for Enhancing Sentiment Analysis of FOMC Minutes. ACM International Conference on AI in Finance","author":"Chen Ziwei","year":"2023","unstructured":"Ziwei Chen, Sandro G\u00f6ssi, Wonseong Kim, Bernhard Bermeitinger, and Siegfried Handschuh. 2023. FinBERT-FOMC: Fine-Tuned FinBERT Model with Sentiment Focus Method for Enhancing Sentiment Analysis of FOMC Minutes. ACM International Conference on AI in Finance (2023), 357\u2013364."},{"key":"e_1_3_2_1_7_1","volume-title":"BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. Conference of the North American","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies 1, Long and Short Papers (2019), 4171\u20134186."},{"key":"e_1_3_2_1_8_1","first-page":"25","article-title":"How You Say It Matters: Text Analysis of FOMC Statements Using Natural Language Processing","volume":"106","author":"Doh Taeyoung","year":"2021","unstructured":"Taeyoung Doh, Sungil Kim, and Shu-Kuei Yang. 2021. How You Say It Matters: Text Analysis of FOMC Statements Using Natural Language Processing. Economic Review-Federal Reserve Bank of Kansas City 106, 1 (2021), 25\u201340.","journal-title":"Economic Review-Federal Reserve Bank of Kansas City"},{"key":"e_1_3_2_1_9_1","first-page":"1877","article-title":"Language Models are Few-Shot Learners","volume":"33","author":"Tom\u00a0Brown","year":"2020","unstructured":"Tom\u00a0Brown et al.2020. Language Models are Few-Shot Learners. Advances in Neural Information Processing Systems 33 (2020), 1877\u20131901.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_10_1","volume-title":"VADER: A Parsimonious Rule-Based Model for Sentiment Analysis of Social Media Text. In International Conference on Weblogs and Social Media (ICWSM-14)","author":"Gilbert Eric","year":"2014","unstructured":"Eric Gilbert. 2014. VADER: A Parsimonious Rule-Based Model for Sentiment Analysis of Social Media Text. In International Conference on Weblogs and Social Media (ICWSM-14), Vol.\u00a08. 216\u2013225."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jmoneco.2019.09.002"},{"key":"e_1_3_2_1_12_1","volume-title":"Proceedings of the 22nd Nordic Conference on Computational Linguistics, Mareike Hartmann and Barbara Plank (Eds.). 187\u2013196","author":"Hoang Mickel","year":"2019","unstructured":"Mickel Hoang, Oskar\u00a0Alija Bihorac, and Jacobo Rouces. 2019. Aspect-Based Sentiment Analysis using BERT. In Proceedings of the 22nd Nordic Conference on Computational Linguistics, Mareike Hartmann and Barbara Plank (Eds.). 187\u2013196."},{"key":"e_1_3_2_1_13_1","unstructured":"Wei Huang Xudong Ma Haotong Qin Xingyu Zheng Chengtao Lv Hong Chen Jie Luo Xiaojuan Qi Xianglong Liu and Michele Magno. 2024. How Good Are Low-bit Quantized LLaMA3 Models? An Empirical Study."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.iref.2020.09.020"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.69473\/iasr.2024.27.1.43"},{"key":"e_1_3_2_1_16_1","first-page":"22199","article-title":"Large Language Models are Zero-Shot Reasoners","volume":"35","author":"Kojima Takeshi","year":"2022","unstructured":"Takeshi Kojima, Shixiang\u00a0Shane Gu, Machel Reid, Yutaka Matsuo, and Yusuke Iwasawa. 2022. Large Language Models are Zero-Shot Reasoners. Advances in Neural Information Processing Systems 35 (2022), 22199\u201322213.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"e_1_3_2_1_18_1","volume-title":"Exploiting BERT for end-to-end aspect-based sentiment analysis. arXiv","author":"Li Xin","year":"2019","unstructured":"Xin Li, Lidong Bing, Wenxuan Zhang, and Wai Lam. 2019. Exploiting BERT for end-to-end aspect-based sentiment analysis. arXiv (2019)."},{"key":"e_1_3_2_1_19_1","volume-title":"Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Industry Track. 408\u2013422","author":"Li Xianzhi","year":"2023","unstructured":"Xianzhi Li, Samuel Chan, Xiaodan Zhu, Yulong Pei, Zhiqiang Ma, Xiaomo Liu, and Sameena Shah. 2023. Are ChatGPT and GPT-4 General-Purpose Solvers for Financial Text Analytics? A Study on Several Typical Tasks. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Industry Track. 408\u2013422."},{"key":"e_1_3_2_1_20_1","volume-title":"RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv (2019)."},{"key":"e_1_3_2_1_21_1","volume-title":"FinBERT: A Pre-Trained Financial Language Representation Model for Financial Text Mining. In International Conference on International Joint Conferences on Artificial Intelligence. 4513\u20134519","author":"Liu Zhuang","year":"2021","unstructured":"Zhuang Liu, Degen Huang, Kaiyu Huang, Zhuang Li, and Jun Zhao. 2021. FinBERT: A Pre-Trained Financial Language Representation Model for Financial Text Mining. In International Conference on International Joint Conferences on Artificial Intelligence. 4513\u20134519."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3184558.3192301"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSC50631.2021.00039"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1002\/asi.23062"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3605943"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3009626"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.finnlp-1.5"},{"key":"e_1_3_2_1_28_1","volume-title":"Deep Contextualized Word Representations. Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT)","author":"Peters Matthew","year":"2018","unstructured":"Matthew Peters, Mark Neumann, Mohit Iyyer\u2020, Matt Gardner\u2020, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. 2018. Deep Contextualized Word Representations. Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT) (2018), 2227\u20132237."},{"key":"e_1_3_2_1_29_1","volume-title":"Improving Language Understanding by Generative Pre-Training. OpenAI","author":"Radford Alex","year":"2018","unstructured":"Alex Radford, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving Language Understanding by Generative Pre-Training. OpenAI (2018)."},{"key":"e_1_3_2_1_30_1","volume-title":"Language Models are Unsupervised Multitask Learners. OpenAI Blog","author":"Radford Alec","year":"2019","unstructured":"Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language Models are Unsupervised Multitask Learners. OpenAI Blog (2019)."},{"key":"e_1_3_2_1_31_1","volume-title":"Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm. Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems","author":"Reynolds Laria","year":"2021","unstructured":"Laria Reynolds and Kyle McDonell. 2021. Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm. Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems (2021), 1\u20137."},{"key":"e_1_3_2_1_32_1","volume-title":"Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering. arXiv","author":"Ridnik Tal","year":"2024","unstructured":"Tal Ridnik, Dedy Kredo, and Itamar Friedman. 2024. Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering. arXiv (2024)."},{"key":"e_1_3_2_1_33_1","unstructured":"Victor Sanh Lysandre Debut Julien Chaumond and Thomas Wolf. 2020. DistilBERT a distilled version of BERT: smaller faster cheaper and lighter."},{"key":"e_1_3_2_1_34_1","volume-title":"Do Monetary Policy Transparency and Central Bank Communication Reduce Interest Rate Disagreement?Journal of Forecasting 39, 3","author":"Seelajaroen Ruttachai","year":"2020","unstructured":"Ruttachai Seelajaroen, Pornanong Budsaratragoon, and Boonlert Jitmaneeroj. 2020. Do Monetary Policy Transparency and Central Bank Communication Reduce Interest Rate Disagreement?Journal of Forecasting 39, 3 (2020), 368\u2013393."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jeconom.2020.07.053"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11633-022-1331-6"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jeconbus.2021.106021"},{"key":"e_1_3_2_1_38_1","volume-title":"Llama: Open and Efficient Foundation Language Models. arXiv","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, 2023. Llama: Open and Efficient Foundation Language Models. arXiv (2023)."},{"key":"e_1_3_2_1_39_1","volume-title":"Advances in Neural Information Processing Systems (NeurIPS) 30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention Is All You Need. Advances in Neural Information Processing Systems (NeurIPS) 30 (2017)."},{"key":"e_1_3_2_1_40_1","volume-title":"Aspect-Based Sentiment Analysis in Document-FOMC Meeting Minutes on Economic Projection. arXiv","author":"Wang Yifei","year":"2021","unstructured":"Yifei Wang. 2021. Aspect-Based Sentiment Analysis in Document-FOMC Meeting Minutes on Economic Projection. arXiv (2021)."},{"key":"e_1_3_2_1_41_1","volume-title":"International Conference on Learning Representations.","author":"Wei Jason","year":"2021","unstructured":"Jason Wei, Maarten Bosma, Vincent Zhao, Kelvin Guu, Adams\u00a0Wei Yu, Brian Lester, Nan Du, Andrew\u00a0M Dai, and Quoc\u00a0V Le. 2021. Finetuned Language Models are Zero-Shot Learners. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_42_1","first-page":"24824","article-title":"Chain-of-Thought Prompting Elicits Reasoning in Large Language Models","volume":"35","author":"Wei Jason","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc Le, and Denny Zhou. 2022. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. Advances in Neural Information Processing Systems 35 (2022), 24824\u201324837.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.3386\/w11898"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.188"}],"event":{"name":"ICAIF '24: 5th ACM International Conference on AI in Finance","location":"Brooklyn NY USA","acronym":"ICAIF '24"},"container-title":["Proceedings of the 5th ACM International Conference on AI in Finance"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3677052.3698675","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3677052.3698675","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T17:10:23Z","timestamp":1755882623000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3677052.3698675"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,14]]},"references-count":44,"alternative-id":["10.1145\/3677052.3698675","10.1145\/3677052"],"URL":"https:\/\/doi.org\/10.1145\/3677052.3698675","relation":{},"subject":[],"published":{"date-parts":[[2024,11,14]]},"assertion":[{"value":"2024-11-14","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}