{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T07:41:52Z","timestamp":1763106112873,"version":"3.45.0"},"publisher-location":"New York, NY, USA","reference-count":41,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,11,15]]},"DOI":"10.1145\/3768292.3770339","type":"proceedings-article","created":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T07:24:26Z","timestamp":1763105066000},"page":"623-631","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Federated Financial Reasoning Distillation: Training A Small Financial Expert by Learning From Multiple Teachers"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-1960-3004","authenticated-orcid":false,"given":"Shuoling","family":"Liu","sequence":"first","affiliation":[{"name":"The Hong Kong University of Science and Technology, Hong Kong, China and E Fund Management Co., Ltd., Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0767-1726","authenticated-orcid":false,"given":"Jiangpeng","family":"Yan","sequence":"additional","affiliation":[{"name":"E Fund Management Co., Ltd., Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-1690-9183","authenticated-orcid":false,"given":"Xiaoyu","family":"Wang","sequence":"additional","affiliation":[{"name":"E Fund Management Co., Ltd., Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-4716-1556","authenticated-orcid":false,"given":"Yuhang","family":"Jiang","sequence":"additional","affiliation":[{"name":"E Fund Management Co., Ltd., Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-9710-9719","authenticated-orcid":false,"given":"Liyuan","family":"Chen","sequence":"additional","affiliation":[{"name":"E Fund Management Co., Ltd., Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-3040-6140","authenticated-orcid":false,"given":"Tao","family":"Fan","sequence":"additional","affiliation":[{"name":"The Hong Kong University of Science and Technology, Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2587-6028","authenticated-orcid":false,"given":"Kai","family":"Chen","sequence":"additional","affiliation":[{"name":"The Hong Kong University of Science and Technology, Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5059-8360","authenticated-orcid":false,"given":"Qiang","family":"Yang","sequence":"additional","affiliation":[{"name":"The Hong Kong Polytechnic University, Hong Kong, China"}]}],"member":"320","published-online":{"date-parts":[[2025,11,14]]},"reference":[{"key":"e_1_3_3_1_2_2","unstructured":"Josh Achiam Steven Adler Sandhini Agarwal Lama Ahmad Ilge Akkaya Florencia\u00a0Leoni Aleman Diogo Almeida Janko Altenschmidt Sam Altman Shyamal Anadkat et\u00a0al. 2023. Gpt-4 technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.08774 (2023)."},{"key":"e_1_3_3_1_3_2","unstructured":"Jinze Bai Shuai Bai Yunfei Chu Zeyu Cui et\u00a0al. 2023. Qwen technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2309.16609 (2023)."},{"key":"e_1_3_3_1_4_2","unstructured":"Yuntao Bai Andy Jones Kamal Ndousse Amanda Askell et\u00a0al. 2022. Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2204.05862 (2022)."},{"key":"e_1_3_3_1_5_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICDMW58026.2022.00078"},{"key":"e_1_3_3_1_6_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i16.29720"},{"key":"e_1_3_3_1_7_2","doi-asserted-by":"crossref","unstructured":"Yupeng Chang Xu Wang Jindong Wang Yuan Wu Linyi Yang Kaijie Zhu Hao Chen Xiaoyuan Yi Cunxiang Wang Yidong Wang et\u00a0al. 2024. A survey on evaluation of large language models. ACM transactions on intelligent systems and technology 15 3 (2024) 1\u201345.","DOI":"10.1145\/3641289"},{"key":"e_1_3_3_1_8_2","doi-asserted-by":"crossref","unstructured":"Tao Fan Hanlin Gu Xuemei Cao Chee\u00a0Seng Chan Qian Chen Yiqiang Chen Yihui Feng et\u00a0al. 2025. Ten challenging problems in federated foundation models. IEEE Transactions on Knowledge and Data Engineering (2025).","DOI":"10.1109\/TKDE.2025.3555328"},{"key":"e_1_3_3_1_9_2","doi-asserted-by":"crossref","unstructured":"Ben Goertzel. 2014. Artificial general intelligence: concept state of the art and future prospects. Journal of Artificial General Intelligence 5 1 (2014) 1.","DOI":"10.2478\/jagi-2014-0001"},{"key":"e_1_3_3_1_10_2","volume-title":"The Twelfth International Conference on Learning Representations","author":"Gu Yuxian","year":"2024","unstructured":"Yuxian Gu, Li Dong, Furu Wei, and Minlie Huang. 2024. MiniLLM: Knowledge Distillation of Large Language Models. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_3_1_11_2","unstructured":"Daya Guo Dejian Yang Haowei Zhang Junxiao Song et\u00a0al. 2025. Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2501.12948 (2025)."},{"key":"e_1_3_3_1_12_2","first-page":"17783","volume-title":"International Conference on Machine Learning","author":"Hayou Soufiane","year":"2024","unstructured":"Soufiane Hayou, Nikhil Ghosh, and Bin Yu. 2024. LoRA+: Efficient Low Rank Adaptation of Large Models. In International Conference on Machine Learning. PMLR, 17783\u201317806."},{"key":"e_1_3_3_1_13_2","unstructured":"Geoffrey Hinton Oriol Vinyals and Jeff Dean. 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/1503.02531 (2015)."},{"key":"e_1_3_3_1_14_2","doi-asserted-by":"crossref","unstructured":"Lei Huang Weijiang Yu Weitao Ma et\u00a0al. 2025. A survey on hallucination in large language models: Principles taxonomy challenges and open questions. ACM Transactions on Information Systems 43 2 (2025) 1\u201355.","DOI":"10.1145\/3703155"},{"key":"e_1_3_3_1_15_2","unstructured":"Aaron Jaech Adam Kalai Adam Lerer Adam Richardson et\u00a0al. 2024. Openai o1 system card. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2412.16720 (2024)."},{"key":"e_1_3_3_1_16_2","unstructured":"Minki Kang Seanie Lee Jinheon Baek Kenji Kawaguchi and Sung\u00a0Ju Hwang. 2023. Knowledge-augmented reasoning distillation for small language models in knowledge-intensive tasks. Advances in Neural Information Processing Systems 36 (2023) 48573\u201348602."},{"key":"e_1_3_3_1_17_2","unstructured":"Takeshi Kojima Shixiang\u00a0Shane Gu Machel Reid Yutaka Matsuo and Yusuke Iwasawa. 2022. Large language models are zero-shot reasoners. Advances in neural information processing systems 35 (2022) 22199\u201322213."},{"key":"e_1_3_3_1_18_2","unstructured":"Weize Kong Spurthi\u00a0Amba Hombaiah Mingyang Zhang Qiaozhu Mei and Michael Bendersky. 2024. Prewrite: Prompt rewriting with reinforcement learning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2401.08189 (2024)."},{"key":"e_1_3_3_1_19_2","unstructured":"Dawei Li Bohan Jiang Liangjie Huang Alimohammad Beigi Chengshuai Zhao Zhen Tan Amrita Bhattacharjee Yuxuan Jiang Canyu Chen Tianhao Wu et\u00a0al. 2024. From generation to judgment: Opportunities and challenges of llm-as-a-judge. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2411.16594 (2024)."},{"key":"e_1_3_3_1_20_2","volume-title":"The Twelfth International Conference on Learning Representations","author":"Lightman Hunter","year":"2023","unstructured":"Hunter Lightman, Vineet Kosaraju, Yuri Burda, et\u00a0al. 2023. Let\u2019s verify step by step. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_3_1_21_2","unstructured":"Ji Lin Jiaming Tang Haotian Tang Shang Yang Wei-Ming Chen Wei-Chen Wang Guangxuan Xiao Xingyu Dang Chuang Gan and Song Han. 2024. Awq: Activation-aware weight quantization for on-device llm compression and acceleration. Proceedings of Machine Learning and Systems 6 (2024) 87\u2013100."},{"key":"e_1_3_3_1_22_2","unstructured":"Aixin Liu Bei Feng Bing Xue Bingxuan\u00a0g Wang et\u00a0al. 2024. Deepseek-v3 technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2412.19437 (2024)."},{"key":"e_1_3_3_1_23_2","unstructured":"Shuoling Liu Liyuan Chen Jiangpeng Yan Yuhang Jiang Xiaoyu Wang Xiu Li and Qiang Yang. 2025. When DeepSeek-R1 meets financial applications: benchmarking opportunities and limitations. Frontiers of Information Technology & Electronic Engineering (2025) 1\u20139."},{"key":"e_1_3_3_1_24_2","unstructured":"Xiao-Yang Liu Guoxuan Wang Hongyang Yang and Daochen Zha. 2023. Fingpt: Democratizing internet-scale data for financial large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2307.10485 (2023)."},{"key":"e_1_3_3_1_25_2","unstructured":"Zhaowei Liu Xin Guo Fangqi Lou et\u00a0al. 2025. Fin-R1: A Large Language Model for Financial Reasoning through Reinforcement Learning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2503.16252 (2025)."},{"key":"e_1_3_3_1_26_2","doi-asserted-by":"crossref","unstructured":"Niklas Muennighoff Zitong Yang Weijia Shi Xiang\u00a0Lisa Li Li Fei-Fei Hannaneh Hajishirzi Luke Zettlemoyer Percy Liang et\u00a0al. 2025. s1: Simple test-time scaling. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2501.19393 (2025).","DOI":"10.18653\/v1\/2025.emnlp-main.1025"},{"key":"e_1_3_3_1_27_2","doi-asserted-by":"crossref","unstructured":"Reid Pryzant Dan Iter Jerry Li Yin\u00a0Tat Lee Chenguang Zhu and Michael Zeng. 2023. Automatic prompt optimization with\" gradient descent\" and beam search. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2305.03495 (2023).","DOI":"10.18653\/v1\/2023.emnlp-main.494"},{"key":"e_1_3_3_1_28_2","doi-asserted-by":"crossref","unstructured":"Kumar Shridhar Alessandro Stolfo and Mrinmaya Sachan. 2023. Distilling reasoning capabilities into smaller language models. Findings of the Association for Computational Linguistics: ACL 2023 (2023) 7059\u20137073.","DOI":"10.18653\/v1\/2023.findings-acl.441"},{"key":"e_1_3_3_1_29_2","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671780"},{"key":"e_1_3_3_1_30_2","first-page":"5878","volume-title":"Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation","author":"Sun Zhihong","year":"2024","unstructured":"Zhihong Sun, Chen Lyu, Bolun Li, , et\u00a0al. 2024. Enhancing Code Generation Performance of Smaller Models by Distilling the Reasoning Ability of LLMs. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation. 5878\u20135895."},{"key":"e_1_3_3_1_31_2","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale et\u00a0al. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2307.09288 (2023)."},{"key":"e_1_3_3_1_32_2","unstructured":"Jason Wei Xuezhi Wang Dale Schuurmans Maarten Bosma Fei Xia et\u00a0al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in neural information processing systems 35 (2022) 24824\u201324837."},{"key":"e_1_3_3_1_33_2","unstructured":"Shijie Wu Ozan Irsoy Steven Lu Vadim Dabravolski Mark Dredze et\u00a0al. 2023. Bloomberggpt: A large language model for finance. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.17564 (2023)."},{"key":"e_1_3_3_1_34_2","unstructured":"Qianqian Xie Weiguang Han Xiao Zhang et\u00a0al. 2023. Pixiu: A comprehensive benchmark instruction dataset and large language model for finance. Advances in Neural Information Processing Systems 36 (2023) 33469\u201333484."},{"key":"e_1_3_3_1_35_2","unstructured":"Xiaohan Xu Ming Li Chongyang Tao Tao Shen Reynold Cheng Jinyang Li Can Xu Dacheng Tao and Tianyi Zhou. 2024. A survey on knowledge distillation of large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.13116 (2024)."},{"key":"e_1_3_3_1_36_2","unstructured":"Yi Yang Yixuan Tang and Kar\u00a0Yan Tam. 2023. Investlm: A large language model for investment using financial domain instruction tuning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2309.13064 (2023)."},{"key":"e_1_3_3_1_37_2","unstructured":"Shunyu Yao Dian Yu Jeffrey Zhao Izhak Shafran et\u00a0al. 2023. Tree of thoughts: Deliberate problem solving with large language models. Advances in neural information processing systems 36 (2023) 11809\u201311822."},{"key":"e_1_3_3_1_38_2","volume-title":"First Conference on Language Modeling","author":"Zelikman Eric","year":"2024","unstructured":"Eric Zelikman, Georges\u00a0Raif Harik, Yijia Shao, Varuna Jayasiri, Nick Haber, and Noah Goodman. 2024. Quiet-star: Language models can teach themselves to think before speaking. In First Conference on Language Modeling."},{"key":"e_1_3_3_1_39_2","unstructured":"Liwen Zhang Weige Cai Zhaowei Liu Zhi Yang et\u00a0al. 2023. Fineval: A chinese financial domain knowledge evaluation benchmark for large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2308.09975 (2023)."},{"key":"e_1_3_3_1_40_2","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3615285"},{"key":"e_1_3_3_1_41_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.acl-demos.38"},{"key":"e_1_3_3_1_42_2","unstructured":"Tianyang Zhong Zhengliang Liu Yi Pan Yutong Zhang et\u00a0al. 2024. Evaluation of openai o1: Opportunities and challenges of agi. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2409.18486 (2024)."}],"event":{"name":"ICAIF '25: 6th ACM International Conference on AI in Finance","location":"Singapore Singapore","acronym":"ICAIF '25"},"container-title":["Proceedings of the 6th ACM International Conference on AI in Finance"],"original-title":[],"deposited":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T07:31:58Z","timestamp":1763105518000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3768292.3770339"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,14]]},"references-count":41,"alternative-id":["10.1145\/3768292.3770339","10.1145\/3768292"],"URL":"https:\/\/doi.org\/10.1145\/3768292.3770339","relation":{},"subject":[],"published":{"date-parts":[[2025,11,14]]},"assertion":[{"value":"2025-11-14","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}