{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T22:12:41Z","timestamp":1773958361625,"version":"3.50.1"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"30","license":[{"start":{"date-parts":[[2024,12,7]],"date-time":"2024-12-07T00:00:00Z","timestamp":1733529600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,7]],"date-time":"2024-12-07T00:00:00Z","timestamp":1733529600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s00521-024-10603-6","type":"journal-article","created":{"date-parts":[[2024,12,7]],"date-time":"2024-12-07T04:00:53Z","timestamp":1733544053000},"page":"24883-24892","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Chinese fine-grained financial sentiment analysis with large language models"],"prefix":"10.1007","volume":"37","author":[{"given":"Yinyu","family":"Lan","sequence":"first","affiliation":[]},{"given":"Yanru","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Wang","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Weiqiang","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Youhao","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,7]]},"reference":[{"issue":"1\u20132","key":"10603_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1561\/1500000011","volume":"2","author":"B Pang","year":"2008","unstructured":"Pang B, Lee L (2008) Opinion mining and sentiment analysis. Found Trends Inf Retr 2(1\u20132):1\u2013135","journal-title":"Found Trends Inf Retr"},{"key":"10603_CR2","doi-asserted-by":"crossref","unstructured":"Cortis K, Freitas A, Daudert T, Huerlimann M, Zarrouk M, Handschuh S, Davis B (2017) Semeval-2017 task 5: fine-grained sentiment analysis on financial microblogs and news. In: Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), pp 519\u2013535","DOI":"10.18653\/v1\/S17-2089"},{"key":"10603_CR3","doi-asserted-by":"publisher","first-page":"272","DOI":"10.1016\/j.eswa.2018.10.003","volume":"118","author":"HH Do","year":"2019","unstructured":"Do HH, Prasad PW, Maag A, Alsadoon A (2019) Deep learning for aspect-based sentiment analysis: a comparative review. Expert Syst Appl 118:272\u2013299","journal-title":"Expert Syst Appl"},{"key":"10603_CR4","doi-asserted-by":"crossref","unstructured":"Maia M, Handschuh S, Freitas A, Davis B, McDermott R, Zarrouk M, Balahur A (2018) Www\u201918 open challenge: financial opinion mining and question answering. In: Companion Proceedings of the the Web Conference 2018, pp 1941\u20131942","DOI":"10.1145\/3184558.3192301"},{"key":"10603_CR5","doi-asserted-by":"publisher","DOI":"10.1016\/j.lindif.2023.102274","volume":"103","author":"E Kasneci","year":"2023","unstructured":"Kasneci E, Se\u00dfler K, K\u00fcchemann S, Bannert M, Dementieva D, Fischer F, Gasser U, Groh G, G\u00fcnnemann S, H\u00fcllermeier E (2023) Chatgpt for good? on opportunities and challenges of large language models for education. Learn Individ Differ 103:102274","journal-title":"Learn Individ Differ"},{"issue":"8","key":"10603_CR6","doi-asserted-by":"publisher","first-page":"1930","DOI":"10.1038\/s41591-023-02448-8","volume":"29","author":"AJ Thirunavukarasu","year":"2023","unstructured":"Thirunavukarasu AJ, Ting DSJ, Elangovan K, Gutierrez L, Tan TF, Ting DSW (2023) Large language models in medicine. Nat Med 29(8):1930\u20131940","journal-title":"Nat Med"},{"key":"10603_CR7","doi-asserted-by":"crossref","unstructured":"Chang Y, Wang X, Wang J, Wu Y, Yang L, Zhu K, Chen H, Yi X, Wang C, Wang Y, et al (2023) A survey on evaluation of large language models. ACM Trans Intell Syst Technol","DOI":"10.1145\/3641289"},{"key":"10603_CR8","unstructured":"Zhao WX, Zhou K, Li J, Tang T, Wang X, Hou Y, Min Y, Zhang B, Zhang J, Dong Z, et al (2023) A survey of large language models. arXiv:2303.18223"},{"key":"10603_CR9","doi-asserted-by":"crossref","unstructured":"Li X, Zhu X, Ma Z, Liu X, Shah S (2023) Are chatgpt and gpt-4 general-purpose solvers for financial text analytics? an examination on several typical tasks. arXiv:2305.05862","DOI":"10.18653\/v1\/2023.emnlp-industry.39"},{"key":"10603_CR10","unstructured":"Touvron H, Lavril T, Izacard G, Martinet X, Lachaux M-A, Lacroix T, Rozi\u00e8re B, Goyal N, Hambro E, Azhar F, et al (2023) Llama: Open and efficient foundation language models. arXiv:2302.13971"},{"key":"10603_CR11","unstructured":"Scao TL, Fan A, Akiki C, Pavlick E, Ili\u0107 S, Hesslow D, Castagn\u00e9 R, Luccioni AS, Yvon F, Gall\u00e9 M, et al (2022) Bloom: A 176b-parameter open-access multilingual language model. arXiv:2211.05100"},{"key":"10603_CR12","unstructured":"Chen Z, Jiang F, Chen J, Wang T, Yu F, Chen G, Zhang H, Liang J, Zhang C, Zhang Z, et al (2023) Phoenix: Democratizing chatgpt across languages. arXiv:2304.10453"},{"key":"10603_CR13","unstructured":"Taori R, Gulrajani I, Zhang T, Dubois Y, Li X, Guestrin C, Liang P, Hashimoto TB (2023) Stanford Alpaca: an instruction-following LLaMA model. GitHub"},{"key":"10603_CR14","doi-asserted-by":"crossref","unstructured":"Wang Y, Kordi Y, Mishra S, Liu A, Smith NA, Khashabi D, Hajishirzi H (2022) Self-instruct: Aligning language model with self generated instructions. arXiv:2212.10560","DOI":"10.18653\/v1\/2023.acl-long.754"},{"key":"10603_CR15","unstructured":"Chiang W-L, Li Z, Lin Z, Sheng Y, Wu Z, Zhang H, Zheng L, Zhuang S, Zhuang Y, Gonzalez JE, Stoica I, Xing EP (2023) Vicuna: an open-source chatbot impressing GPT-4 with 90%* ChatGPT Quality. https:\/\/lmsys.org\/blog\/2023-03-30-vicuna\/"},{"key":"10603_CR16","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser, \u0141, Polosukhin I (2017) Attention is all you need. Advances in Neural Information Processing Systems 30"},{"key":"10603_CR17","unstructured":"Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805"},{"key":"10603_CR18","unstructured":"Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Levy O, Lewis M, Zettlemoyer L, Stoyanov V (2019) Roberta: a robustly optimized bert pretraining approach. arXiv:1907.11692"},{"key":"10603_CR19","unstructured":"Lan Z, Chen M, Goodman S, Gimpel K, Sharma P, Soricut R (2019) Albert: a lite bert for self-supervised learning of language representations. arXiv:1909.11942"},{"key":"10603_CR20","doi-asserted-by":"crossref","unstructured":"Karl F, Scherp A (2022) Transformers are short text classifiers: a study of inductive short text classifiers on benchmarks and real-world datasets. arXiv:2211.16878","DOI":"10.1007\/978-3-031-40837-3_7"},{"key":"10603_CR21","first-page":"37309","volume":"35","author":"M Yasunaga","year":"2022","unstructured":"Yasunaga M, Bosselut A, Ren H, Zhang X, Manning CD, Liang PS, Leskovec J (2022) Deep bidirectional language-knowledge graph pretraining. Adv Neural Inf Process Syst 35:37309\u201337323","journal-title":"Adv Neural Inf Process Syst"},{"key":"10603_CR22","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s12911-020-01362-0","volume":"21","author":"Y Lan","year":"2021","unstructured":"Lan Y, He S, Liu K, Zeng X, Liu S, Zhao J (2021) Path-based knowledge reasoning with textual semantic information for medical knowledge graph completion. BMC Med Inf Decis Mak 21:1\u201312","journal-title":"BMC Med Inf Decis Mak"},{"key":"10603_CR23","doi-asserted-by":"crossref","unstructured":"Wang Y, Zhang H, Liang J, Li R (2023) Dynamic heterogeneous-graph reasoning with language models and knowledge representation learning for commonsense question answering. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 14048\u201314063","DOI":"10.18653\/v1\/2023.acl-long.785"},{"key":"10603_CR24","unstructured":"Radford A, Narasimhan K, Salimans T, Sutskever I, et al (2018) Improving language understanding by generative pre-training"},{"issue":"8","key":"10603_CR25","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I (2019) Language models are unsupervised multitask learners. OpenAI blog 1(8):9","journal-title":"OpenAI blog"},{"issue":"1","key":"10603_CR26","first-page":"5485","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel C, Shazeer N, Roberts A, Lee K, Narang S, Matena M, Zhou Y, Li W, Liu PJ (2020) Exploring the limits of transfer learning with a unified text-to-text transformer. J Mach Learn Res 21(1):5485\u20135551","journal-title":"J Mach Learn Res"},{"key":"10603_CR27","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A (2020) Language models are few-shot learners. Adv Neural Inf Process Syst 33:1877\u20131901","journal-title":"Adv Neural Inf Process Syst"},{"key":"10603_CR28","unstructured":"Wu S, Irsoy O, Lu S, Dabravolski V, Dredze M, Gehrmann S, Kambadur P, Rosenberg D, Mann G (2023) Bloomberggpt: a large language model for finance. arXiv:2303.17564"},{"key":"10603_CR29","doi-asserted-by":"crossref","unstructured":"Yang H, Liu X-Y, Wang CD (2023) Fingpt: open-source financial large language models. arXiv:2306.06031","DOI":"10.2139\/ssrn.4489826"},{"key":"10603_CR30","first-page":"27730","volume":"35","author":"L Ouyang","year":"2022","unstructured":"Ouyang L, Wu J, Jiang X, Almeida D, Wainwright C, Mishkin P, Zhang C, Agarwal S, Slama K, Ray A (2022) Training language models to follow instructions with human feedback. Adv Neural Inf Process Syst 35:27730\u201327744","journal-title":"Adv Neural Inf Process Syst"},{"key":"10603_CR31","doi-asserted-by":"crossref","unstructured":"Lu Y, Liu Q, Dai D, Xiao X, Lin H, Han X, Sun L, Wu H (2022) Unified structure generation for universal information extraction. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 55, pp 5755\u20135772","DOI":"10.18653\/v1\/2022.acl-long.395"},{"key":"10603_CR32","unstructured":"Beltagy I, Peters ME, Cohan A (2020) Longformer: the long-document transformer. arXiv:2004.05150"},{"key":"10603_CR33","unstructured":"Cui Y, Yang Z, Yao X (2023) Efficient and effective text encoding for Chinese llama and alpaca. arXiv:2304.08177"},{"key":"10603_CR34","doi-asserted-by":"crossref","unstructured":"Muennighoff N, Wang T, Sutawika L, Roberts A, Biderman S, Scao TL, Bari MS, Shen S, Yong ZX, Schoelkopf H, et al (2022) Crosslingual generalization through multitask finetuning. arXiv:2211.01786","DOI":"10.18653\/v1\/2023.acl-long.891"},{"key":"10603_CR35","doi-asserted-by":"crossref","unstructured":"Du Z, Qian Y, Liu X, Ding M, Qiu J, Yang Z, Tang J (2022) Glm: general language model pretraining with autoregressive blank infilling. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 320\u2013335","DOI":"10.18653\/v1\/2022.acl-long.26"},{"key":"10603_CR36","unstructured":"Zeng A, Liu X, Du Z, Wang Z, Lai H, Ding M, Yang Z, Xu Y, Zheng W, Xia X, et al (2022) Glm-130b: an open bilingual pre-trained model. arXiv:2210.02414"},{"key":"10603_CR37","unstructured":"Wei J, Bosma M, Zhao VY, Guu K, Yu AW, Lester B, Du N, Dai AM, Le QV (2021) Finetuned language models are zero-shot learners. arXiv:2109.01652"},{"key":"10603_CR38","doi-asserted-by":"crossref","unstructured":"Rajbhandari S, Rasley J, Ruwase O, He Y (2020) Zero: memory optimizations toward training trillion parameter models. In: SC20: International Conference for High Performance Computing, Networking, Storage and Analysis, IEEE, pp 1\u201316","DOI":"10.1109\/SC41405.2020.00024"},{"key":"10603_CR39","doi-asserted-by":"crossref","unstructured":"Zhao Y, Gu A, Varma R, Luo, L, Huang C-C, Xu M, Wright L, Shojanazeri H, Ott M, Shleifer S, et al (2023) Pytorch fsdp: experiences on scaling fully sharded data parallel. arXiv:2304.11277","DOI":"10.14778\/3611540.3611569"},{"key":"10603_CR40","doi-asserted-by":"crossref","unstructured":"Liu X, Ji K, Fu Y, Tam WL, Du Z, Yang Z, Tang J (2021) P-tuning v2: prompt tuning can be comparable to fine-tuning universally across scales and tasks. arXiv:2110.07602","DOI":"10.18653\/v1\/2022.acl-short.8"},{"key":"10603_CR41","unstructured":"Hu EJ, Shen Y, Wallis P, Allen-Zhu Z, Li Y, Wang S, Wang L, Chen W (2021) Lora: low-rank adaptation of large language models. arXiv:2106.09685"},{"key":"10603_CR42","unstructured":"Gu A, Dao T (2023) Mamba: linear-time sequence modeling with selective state spaces. arXiv:2312.00752"},{"key":"10603_CR43","unstructured":"Dao T, Gu A (2024) Transformers are ssms: generalized models and efficient algorithms through structured state space duality. arXiv:2405.21060"},{"key":"10603_CR44","doi-asserted-by":"crossref","unstructured":"Peng B, Alcaide E, Anthony Q, Albalak A, Arcadinho S, Biderman S, Cao H, Cheng X, Chung M, Grella M, et al (2023) RWKV: reinventing RNNS for the transformer era. arXiv:2305.13048","DOI":"10.18653\/v1\/2023.findings-emnlp.936"},{"key":"10603_CR45","unstructured":"Beck M, P\u00f6ppel K, Spanring M, Auer A, Prudnikova O, Kopp M, Klambauer G, Brandstetter J, Hochreiter S (2024) XLSTM: extended long short-term memory. arXiv:2405.04517"},{"key":"10603_CR46","unstructured":"Mercat J, Vasiljevic I, Keh S, Arora K, Dave A, Gaidon A, Kollar T (2024) Linearizing large language models. arXiv:2405.06640"},{"key":"10603_CR47","unstructured":"Yang S, Wang B, Zhang Y, Shen Y, Kim Y (2024) Parallelizing linear transformers with the delta rule over sequence length. arXiv:2406.06484"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-024-10603-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-024-10603-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-024-10603-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T12:36:55Z","timestamp":1760013415000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-024-10603-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,7]]},"references-count":47,"journal-issue":{"issue":"30","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["10603"],"URL":"https:\/\/doi.org\/10.1007\/s00521-024-10603-6","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"value":"0941-0643","type":"print"},{"value":"1433-3058","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,7]]},"assertion":[{"value":"3 December 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 October 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 December 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declared that they have no conflict of interest in this work.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}