{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T23:21:58Z","timestamp":1773357718870,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":36,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,11,14]],"date-time":"2024-11-14T00:00:00Z","timestamp":1731542400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100006374","name":"Vlaamse Overheid","doi-asserted-by":"publisher","award":["3E240430"],"award-info":[{"award-number":["3E240430"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100006374","name":"Agentschap Innoveren en Ondernemen","doi-asserted-by":"publisher","award":["HBC.2020.2883"],"award-info":[{"award-number":["HBC.2020.2883"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,11,14]]},"DOI":"10.1145\/3677052.3698628","type":"proceedings-article","created":{"date-parts":[[2024,11,14]],"date-time":"2024-11-14T06:38:06Z","timestamp":1731566286000},"page":"283-291","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["A Dutch Financial Large Language Model"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6042-8461","authenticated-orcid":false,"given":"Sander","family":"Noels","sequence":"first","affiliation":[{"name":"Department of Electronics and Information Systems, Ghent University, Belgium and Silverfin, Belgium"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-5911-9703","authenticated-orcid":false,"given":"Jorne","family":"De Blaere","sequence":"additional","affiliation":[{"name":"Silverfin, Belgium"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2692-7504","authenticated-orcid":false,"given":"Tijl","family":"De Bie","sequence":"additional","affiliation":[{"name":"Department of Electronics and Information Systems, Ghent University, Belgium"}]}],"member":"320","published-online":{"date-parts":[[2024,11,14]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Proceedings of the Australasian Language Technology Association Workshop","author":"Cesar\u00a0Salinas Alvarado Julio","year":"2015","unstructured":"Julio Cesar\u00a0Salinas Alvarado, Karin Verspoor, and Timothy Baldwin. 2015. Domain adaption of named entity recognition to support credit risk assessment. In Proceedings of the Australasian Language Technology Association Workshop 2015. 84\u201390."},{"key":"e_1_3_2_1_2_1","volume-title":"Finbert: Financial sentiment analysis with pre-trained language models. arXiv preprint arXiv:1908.10063","author":"Araci Dogu","year":"2019","unstructured":"Dogu Araci. 2019. Finbert: Financial sentiment analysis with pre-trained language models. arXiv preprint arXiv:1908.10063 (2019)."},{"key":"e_1_3_2_1_3_1","volume-title":"FinTral: A Family of GPT-4 Level Multimodal Financial Large Language Models. arXiv preprint arXiv:2402.10986","author":"Bhatia Gagan","year":"2024","unstructured":"Gagan Bhatia, El\u00a0Moatez\u00a0Billah Nagoudi, Hasan Cavusoglu, and Muhammad Abdul-Mageed. 2024. FinTral: A Family of GPT-4 Level Multimodal Financial Large Language Models. arXiv preprint arXiv:2402.10986 (2024)."},{"key":"e_1_3_2_1_4_1","volume-title":"Language models are few-shot learners. Advances in neural information processing systems 33","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared\u00a0D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, 2020. Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877\u20131901."},{"key":"e_1_3_2_1_5_1","volume-title":"Convfinqa: Exploring the chain of numerical reasoning in conversational finance question answering. arXiv preprint arXiv:2210.03849","author":"Chen Zhiyu","year":"2022","unstructured":"Zhiyu Chen, Shiyang Li, Charese Smiley, Zhiqiang Ma, Sameena Shah, and William\u00a0Yang Wang. 2022. Convfinqa: Exploring the chain of numerical reasoning in conversational finance question answering. arXiv preprint arXiv:2210.03849 (2022)."},{"key":"e_1_3_2_1_6_1","first-page":"16344","article-title":"Flashattention: Fast and memory-efficient exact attention with io-awareness","volume":"35","author":"Dao Tri","year":"2022","unstructured":"Tri Dao, Dan Fu, Stefano Ermon, Atri Rudra, and Christopher R\u00e9. 2022. Flashattention: Fast and memory-efficient exact attention with io-awareness. Advances in Neural Information Processing Systems 35 (2022), 16344\u201316359.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543873.3587324"},{"key":"e_1_3_2_1_8_1","volume-title":"Qlora: Efficient finetuning of quantized llms. Advances in Neural Information Processing Systems 36","author":"Dettmers Tim","year":"2024","unstructured":"Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, and Luke Zettlemoyer. 2024. Qlora: Efficient finetuning of quantized llms. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_2_1_9_1","volume-title":"Construction of a japanese financial benchmark for large language models. arXiv preprint arXiv:2403.15062","author":"Hirano Masanori","year":"2024","unstructured":"Masanori Hirano. 2024. Construction of a japanese financial benchmark for large language models. arXiv preprint arXiv:2403.15062 (2024)."},{"key":"e_1_3_2_1_10_1","volume-title":"Instruction Data, and Benchmarks. arXiv preprint arXiv:2403.06249","author":"Hu Gang","year":"2024","unstructured":"Gang Hu, Ke Qin, Chenhan Yuan, Min Peng, Alejandro Lopez-Lira, Benyou Wang, Sophia Ananiadou, Wanlong Yu, Jimin Huang, and Qianqian Xie. 2024. No Language is an Island: Unifying Chinese and English in Financial Large Language Models, Instruction Data, and Benchmarks. arXiv preprint arXiv:2403.06249 (2024)."},{"key":"e_1_3_2_1_11_1","volume-title":"Florian Bressand, Gianna Lengyel","author":"Jiang Q","year":"2023","unstructured":"Albert\u00a0Q Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra\u00a0Singh Chaplot, Diego de\u00a0las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lucile Saulnier, 2023. Mistral 7B. arXiv preprint arXiv:2310.06825 (2023)."},{"key":"e_1_3_2_1_12_1","volume-title":"Proceedings of the 1st Joint Workshop on Financial Narrative Processing and MultiLing Financial Summarisation. 118\u2013123","author":"La\u00a0Quatra Moreno","year":"2020","unstructured":"Moreno La\u00a0Quatra and Luca Cagliero. 2020. End-to-end training for financial report summarization. In Proceedings of the 1st Joint Workshop on Financial Narrative Processing and MultiLing Financial Summarisation. 118\u2013123."},{"key":"e_1_3_2_1_13_1","volume-title":"Okapi: Instruction-tuned large language models in multiple languages with reinforcement learning from human feedback. arXiv preprint arXiv:2307.16039","author":"Lai Viet\u00a0Dac","year":"2023","unstructured":"Viet\u00a0Dac Lai, Chien Van\u00a0Nguyen, Nghia\u00a0Trung Ngo, Thuat Nguyen, Franck Dernoncourt, Ryan\u00a0A Rossi, and Thien\u00a0Huu Nguyen. 2023. Okapi: Instruction-tuned large language models in multiple languages with reinforcement learning from human feedback. arXiv preprint arXiv:2307.16039 (2023)."},{"key":"e_1_3_2_1_14_1","volume-title":"Bbt-fin: Comprehensive construction of chinese financial domain pre-trained language model, corpus and benchmark. arXiv preprint arXiv:2302.09432","author":"Lu Dakuan","year":"2023","unstructured":"Dakuan Lu, Hengkui Wu, Jiaqing Liang, Yipei Xu, Qianyu He, Yipeng Geng, Mengkun Han, Yingsi Xin, and Yanghua Xiao. 2023. Bbt-fin: Comprehensive construction of chinese financial domain pre-trained language model, corpus and benchmark. arXiv preprint arXiv:2302.09432 (2023)."},{"key":"e_1_3_2_1_15_1","unstructured":"Neural Magic. 2022. Twitter financial news sentiment."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3184558.3192301"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1002\/asi.23062"},{"key":"e_1_3_2_1_18_1","volume-title":"Cross-task generalization via natural language crowdsourcing instructions. arXiv preprint arXiv:2104.08773","author":"Mishra Swaroop","year":"2021","unstructured":"Swaroop Mishra, Daniel Khashabi, Chitta Baral, and Hannaneh Hajishirzi. 2021. Cross-task generalization via natural language crowdsourcing instructions. arXiv preprint arXiv:2104.08773 (2021)."},{"key":"e_1_3_2_1_19_1","volume-title":"A Survey of Large Language Models for Financial Applications: Progress, Prospects and Challenges. arXiv preprint arXiv:2406.11903","author":"Nie Yuqi","year":"2024","unstructured":"Yuqi Nie, Yaxuan Kong, Xiaowen Dong, John\u00a0M Mulvey, H\u00a0Vincent Poor, Qingsong Wen, and Stefan Zohren. 2024. A Survey of Large Language Models for Financial Applications: Progress, Prospects and Challenges. arXiv preprint arXiv:2406.11903 (2024)."},{"key":"e_1_3_2_1_20_1","volume-title":"Machine Learning and Knowledge Discovery in Databases: Applied Data Science and Demo Track, Gianmarco De\u00a0Francisci\u00a0Morales","author":"Noels Sander","unstructured":"Sander Noels, Adriaan Merlevede, Andrew Fecheyr, Maarten Vanhalst, Nick Meerlaen, S\u00e9bastien Viaene, and Tijl De\u00a0Bie. 2023. Automated Financial Analysis Using GPT-4. In Machine Learning and Knowledge Discovery in Databases: Applied Data Science and Demo Track, Gianmarco De\u00a0Francisci\u00a0Morales, Claudia Perlich, Natali Ruchansky, Nicolas Kourtellis, Elena Baralis, and Francesco Bonchi (Eds.). Springer Nature Switzerland, Cham, 345\u2013349."},{"key":"e_1_3_2_1_21_1","unstructured":"Sander Noels S\u00e9bastien Viaene and Tijl\u00a0De Bie. 2024. TopoLedgerBERT: Topological Learning of Ledger Description Embeddings using Siamese BERT-Networks. arxiv:2407.05175\u00a0[cs.CE] https:\/\/arxiv.org\/abs\/2407.05175"},{"key":"e_1_3_2_1_22_1","volume-title":"Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems 36","author":"Rafailov Rafael","year":"2024","unstructured":"Rafael Rafailov, Archit Sharma, Eric Mitchell, Christopher\u00a0D Manning, Stefano Ermon, and Chelsea Finn. 2024. Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"crossref","unstructured":"Laria Reynolds and Kyle McDonell. 2021. Prompt programming for large language models: Beyond the few-shot paradigm. In Extended abstracts of the 2021 CHI conference on human factors in computing systems. 1\u20137.","DOI":"10.1145\/3411763.3451760"},{"key":"e_1_3_2_1_24_1","volume-title":"Multitask prompted training enables zero-shot task generalization. arXiv preprint arXiv:2110.08207","author":"Sanh Victor","year":"2021","unstructured":"Victor Sanh, Albert Webson, Colin Raffel, Stephen\u00a0H Bach, Lintang Sutawika, Zaid Alyafeai, Antoine Chaffin, Arnaud Stiegler, Teven\u00a0Le Scao, Arun Raja, 2021. Multitask prompted training enables zero-shot task generalization. arXiv preprint arXiv:2110.08207 (2021)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3487553.3524637"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-73103-8_41"},{"key":"e_1_3_2_1_27_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale Dan Bikel Lukas Blecher Cristian\u00a0Canton Ferrer Moya Chen Guillem Cucurull David Esiobu Jude Fernandes Jeremy Fu Wenyin Fu Brian Fuller Cynthia Gao Vedanuj Goswami Naman Goyal Anthony Hartshorn Saghar Hosseini Rui Hou Hakan Inan Marcin Kardas Viktor Kerkez Madian Khabsa Isabel Kloumann Artem Korenev Punit\u00a0Singh Koura Marie-Anne Lachaux Thibaut Lavril Jenya Lee Diana Liskovich Yinghai Lu Yuning Mao Xavier Martinet Todor Mihaylov Pushkar Mishra Igor Molybog Yixin Nie Andrew Poulton Jeremy Reizenstein Rashi Rungta Kalyan Saladi Alan Schelten Ruan Silva Eric\u00a0Michael Smith Ranjan Subramanian Xiaoqing\u00a0Ellen Tan Binh Tang Ross Taylor Adina Williams Jian\u00a0Xiang Kuan Puxin Xu Zheng Yan Iliyan Zarov Yuchen Zhang Angela Fan Melanie Kambadur Sharan Narang Aurelien Rodriguez Robert Stojnic Sergey Edunov and Thomas Scialom. 2023. Llama 2: Open Foundation and Fine-Tuned Chat Models. arxiv:2307.09288\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2307.09288"},{"key":"e_1_3_2_1_28_1","volume-title":"Language resources for Dutch large language modelling. arXiv preprint arXiv:2312.12852","author":"Vanroy Bram","year":"2023","unstructured":"Bram Vanroy. 2023. Language resources for Dutch large language modelling. arXiv preprint arXiv:2312.12852 (2023)."},{"key":"e_1_3_2_1_29_1","volume-title":"Fingpt: Instruction tuning benchmark for open-source large language models in financial datasets. arXiv preprint arXiv:2310.04793","author":"Wang Neng","year":"2023","unstructured":"Neng Wang, Hongyang Yang, and Christina\u00a0Dan Wang. 2023. Fingpt: Instruction tuning benchmark for open-source large language models in financial datasets. arXiv preprint arXiv:2310.04793 (2023)."},{"key":"e_1_3_2_1_30_1","volume-title":"Self-instruct: Aligning language models with self-generated instructions. arXiv preprint arXiv:2212.10560","author":"Wang Yizhong","year":"2022","unstructured":"Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Alisa Liu, Noah\u00a0A Smith, Daniel Khashabi, and Hannaneh Hajishirzi. 2022. Self-instruct: Aligning language models with self-generated instructions. arXiv preprint arXiv:2212.10560 (2022)."},{"key":"e_1_3_2_1_31_1","volume-title":"Finetuned language models are zero-shot learners. arXiv preprint arXiv:2109.01652","author":"Wei Jason","year":"2021","unstructured":"Jason Wei, Maarten Bosma, Vincent\u00a0Y Zhao, Kelvin Guu, Adams\u00a0Wei Yu, Brian Lester, Nan Du, Andrew\u00a0M Dai, and Quoc\u00a0V Le. 2021. Finetuned language models are zero-shot learners. arXiv preprint arXiv:2109.01652 (2021)."},{"key":"e_1_3_2_1_32_1","volume-title":"Bloomberggpt: A large language model for finance. arXiv preprint arXiv:2303.17564","author":"Wu Shijie","year":"2023","unstructured":"Shijie Wu, Ozan Irsoy, Steven Lu, Vadim Dabravolski, Mark Dredze, Sebastian Gehrmann, Prabhanjan Kambadur, David Rosenberg, and Gideon Mann. 2023. Bloomberggpt: A large language model for finance. arXiv preprint arXiv:2303.17564 (2023)."},{"key":"e_1_3_2_1_33_1","volume-title":"Pixiu: A large language model, instruction data and evaluation benchmark for finance. arXiv preprint arXiv:2306.05443","author":"Xie Qianqian","year":"2023","unstructured":"Qianqian Xie, Weiguang Han, Xiao Zhang, Yanzhao Lai, Min Peng, Alejandro Lopez-Lira, and Jimin Huang. 2023. Pixiu: A large language model, instruction data and evaluation benchmark for finance. arXiv preprint arXiv:2306.05443 (2023)."},{"key":"e_1_3_2_1_34_1","volume-title":"Fingpt: Open-source financial large language models. arXiv preprint arXiv:2306.06031","author":"Yang Hongyang","year":"2023","unstructured":"Hongyang Yang, Xiao-Yang Liu, and Christina\u00a0Dan Wang. 2023. Fingpt: Open-source financial large language models. arXiv preprint arXiv:2306.06031 (2023)."},{"key":"e_1_3_2_1_35_1","volume-title":"arXiv preprint arXiv:2402.07405","author":"Zhang Xiao","year":"2024","unstructured":"Xiao Zhang, Ruoyu Xiang, Chenhan Yuan, Duanyu Feng, Weiguang Han, Alejandro Lopez-Lira, Xiao-Yang Liu, Sophia Ananiadou, Min Peng, Jimin Huang, 2024. D\\ \u2019olares or Dollars? Unraveling the Bilingual Prowess of Financial LLMs Between Spanish and English. arXiv preprint arXiv:2402.07405 (2024)."},{"key":"e_1_3_2_1_36_1","volume-title":"Mengzi: Towards lightweight yet ingenious pre-trained models for chinese. arXiv preprint arXiv:2110.06696","author":"Zhang Zhuosheng","year":"2021","unstructured":"Zhuosheng Zhang, Hanqing Zhang, Keming Chen, Yuhang Guo, Jingyun Hua, Yulong Wang, and Ming Zhou. 2021. Mengzi: Towards lightweight yet ingenious pre-trained models for chinese. arXiv preprint arXiv:2110.06696 (2021)."}],"event":{"name":"ICAIF '24: 5th ACM International Conference on AI in Finance","location":"Brooklyn NY USA","acronym":"ICAIF '24"},"container-title":["Proceedings of the 5th ACM International Conference on AI in Finance"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3677052.3698628","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3677052.3698628","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T17:10:55Z","timestamp":1755882655000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3677052.3698628"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,14]]},"references-count":36,"alternative-id":["10.1145\/3677052.3698628","10.1145\/3677052"],"URL":"https:\/\/doi.org\/10.1145\/3677052.3698628","relation":{},"subject":[],"published":{"date-parts":[[2024,11,14]]},"assertion":[{"value":"2024-11-14","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}