{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T15:28:17Z","timestamp":1772119697240,"version":"3.50.1"},"reference-count":48,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,8,2]],"date-time":"2025-08-02T00:00:00Z","timestamp":1754092800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,8,2]],"date-time":"2025-08-02T00:00:00Z","timestamp":1754092800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Program of Science and Technol- ogy Development Plan of Jilin Province of China","award":["20240101374JC"],"award-info":[{"award-number":["20240101374JC"]}]},{"name":"Shenzhen Science and Technology Program","award":["JCYJ20230807150300001"],"award-info":[{"award-number":["JCYJ20230807150300001"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Intell Inf Syst"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s10844-025-00971-3","type":"journal-article","created":{"date-parts":[[2025,8,2]],"date-time":"2025-08-02T06:37:33Z","timestamp":1754116653000},"page":"1955-1980","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["From news to trends: a financial time series forecasting framework with LLM-driven news sentiment analysis and selective state spaces"],"prefix":"10.1007","volume":"63","author":[{"given":"Renjie","family":"Wang","sequence":"first","affiliation":[]},{"given":"Minghui","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Limin","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,2]]},"reference":[{"key":"971_CR1","doi-asserted-by":"publisher","unstructured":"Ainslie, J., Lee-Thorp, J., de\u00a0Jong, M., et\u00a0al. (2023). Gqa: Training generalized multi-query transformer models from multi-head checkpoints. In: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pp 4895\u20134901, https:\/\/doi.org\/10.18653\/v1\/2023.emnlp-main.298","DOI":"10.18653\/v1\/2023.emnlp-main.298"},{"key":"971_CR2","unstructured":"Anthony, Q., Tokpanov, Y., Glorioso, P., et\u00a0al. (2024). Blackmamba: Mixture of experts for state-space models. arXiv:2402.01771"},{"key":"971_CR3","doi-asserted-by":"publisher","unstructured":"Chen, Q., Kawashima, H. (2024). Stock price prediction using llm-based sentiment analysis. In: 2024 IEEE International Conference on Big Data (BigData), pp. 4846\u20134853, https:\/\/doi.org\/10.1109\/BigData62323.2024.10825946","DOI":"10.1109\/BigData62323.2024.10825946"},{"key":"971_CR4","unstructured":"Chen, S.A., Li, C.L., Arik, S.O., et\u00a0al. (2023). TSMixer: An all-MLP architecture for time series forecast-ing. Transactions on Machine Learning Research https:\/\/openreview.net\/forum?id=wbpxTuXgm0"},{"key":"971_CR5","doi-asserted-by":"publisher","unstructured":"Cheng, D., Yang, F., Xiang, S., et\u00a0al. (2022). Financial time series forecasting with multi-modality graph neural network. Pattern Recognition 121, 108218. https:\/\/doi.org\/10.1016\/j.patcog.2021.108218","DOI":"10.1016\/j.patcog.2021.108218"},{"key":"971_CR6","doi-asserted-by":"crossref","unstructured":"Cho, K., van Merrienboer, B., Gulcehre, C., et\u00a0al. (2014). Learning phrase representations using rnn encoder-decoder for statistical machine translation. In: Conference on Empirical Methods in Natural Language Processing (EMNLP 2014), arXiv:1406.1078","DOI":"10.3115\/v1\/D14-1179"},{"key":"971_CR7","doi-asserted-by":"crossref","unstructured":"Dai, D., Deng, C., Zhao, C., et\u00a0al. (2024). Deepseekmoe: Towards ultimate expert specialization in mixture-of-experts language models. In: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics, pp. 1280\u20131297, arXiv:2401.06066","DOI":"10.18653\/v1\/2024.acl-long.70"},{"key":"971_CR8","unstructured":"DeepSeek-AI,., Guo, D., Yang, D., et\u00a0al. (2025). Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. arXiv:2501.12948"},{"key":"971_CR9","doi-asserted-by":"publisher","unstructured":"Devlin, J., Chang, M.W., Lee, K., et\u00a0al. (2019). Bert: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4171\u20134186, https:\/\/doi.org\/10.18653\/v1\/N19-1423","DOI":"10.18653\/v1\/N19-1423"},{"key":"971_CR10","doi-asserted-by":"publisher","unstructured":"Dosovitskiy, A. (2020). An image is worth 16x16 words: Transformers for image recognition at scale. In: International Conference on Learning Representations, https:\/\/doi.org\/10.48550\/arXiv.2010.11929","DOI":"10.48550\/arXiv.2010.11929"},{"key":"971_CR11","doi-asserted-by":"publisher","unstructured":"Ekambaram, V., Jati, A., Nguyen, N., et\u00a0al. (2023). Tsmixer: Lightweight mlp-mixer model for multivariate time series forecasting. In: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 459\u2013469, https:\/\/doi.org\/10.48550\/arXiv.2306.09364","DOI":"10.48550\/arXiv.2306.09364"},{"key":"971_CR12","doi-asserted-by":"publisher","unstructured":"El\u00a0Zaar, A., Mansouri, A., Benaya, N., et\u00a0al. (2025). Hybrid transformer-cnn architecture for multivariate time series forecasting: Integrating attention mechanisms with convolutional feature extraction. Journal of Intelligent Information Systems pp. 1\u201332. https:\/\/doi.org\/10.1007\/s10844-025-00937-5","DOI":"10.1007\/s10844-025-00937-5"},{"key":"971_CR13","doi-asserted-by":"crossref","unstructured":"Elfwing, S., Uchibe, E., Doya, K. (2018). Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Networks 107, 3\u201311. arXiv:1702.03118","DOI":"10.1016\/j.neunet.2017.12.012"},{"key":"971_CR14","doi-asserted-by":"crossref","unstructured":"Feng, F., Chen, H., He, X., et\u00a0al. (2019). Enhancing stock movement prediction with adversarial training. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19), https:\/\/www.ijcai.org\/proceedings\/2019\/0810.pdf","DOI":"10.24963\/ijcai.2019\/810"},{"key":"971_CR15","doi-asserted-by":"publisher","unstructured":"Gu, A., Dao, T. (2023). Mamba: Linear-time sequence modeling with selective state spaces. https:\/\/doi.org\/10.48550\/arXiv.2312.00752","DOI":"10.48550\/arXiv.2312.00752"},{"key":"971_CR16","unstructured":"Gu, A., Dao, T., Ermon, S., et\u00a0al. (2020). Hippo: Recurrent memory with optimal polynomial projections. In: Larochelle, H., Ranzato, M., Hadsell, R., et\u00a0al. (eds.) Advances in Neural Information Processing Systems, 33. Curran Associates, Inc., pp. 1474\u20131487, arXiv:2008.07669"},{"key":"971_CR17","unstructured":"Gu, A., Goel, K., R\u00e9, C. (2022). Efficiently modeling long sequences with structured state spaces. In: International Conference on Learning Representations, arXiv:2111.00396"},{"issue":"8","key":"971_CR18","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735\u20131780. https:\/\/doi.org\/10.1162\/neco.1997.9.8.1735","journal-title":"Neural Computation"},{"key":"971_CR19","unstructured":"Hu, E.J., Shen, Y., Wallis, P., et\u00a0al. (2022). Lora: Low-rank adaptation of large language models. In: International Conference on Learning Representations, p\u00a03, https:\/\/openreview.net\/forum?id=nZeVKeeFYf9"},{"key":"971_CR20","doi-asserted-by":"publisher","unstructured":"Huang, A.H., Wang, H., Yang, Y. (2023). Finbert: A large language model for extracting information from financial text. Contemporary Accounting Research 40(2), 806\u2013841. https:\/\/doi.org\/10.1111\/1911-3846.12832","DOI":"10.1111\/1911-3846.12832"},{"key":"971_CR21","doi-asserted-by":"publisher","unstructured":"Inserte, P.R., Nakhl\u00e9, M., Qader, R., et\u00a0al. (2023). Large language model adaptation for financial sentiment analysis. In: Proceedings of the Sixth Workshop on Financial Technology and Natural Language Processing, pp. 1\u201310, https:\/\/doi.org\/10.18653\/v1\/2023.finnlp-2.1","DOI":"10.18653\/v1\/2023.finnlp-2.1"},{"key":"971_CR22","doi-asserted-by":"crossref","unstructured":"Kim, Y. (2014). Convolutional neural networks for sentence classification. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1746\u20131751, arXiv:1408.5882","DOI":"10.3115\/v1\/D14-1181"},{"key":"971_CR23","doi-asserted-by":"publisher","unstructured":"Liu, H., Zhang, Y., Wang, X., et\u00a0al. (2023). St-moe: Spatio-temporal mixture of experts for multivariate time series forecasting. In: 2023 18th International Conference on Intelligent Systems and Knowledge Engineering (ISKE), pp. 562\u2013567, https:\/\/doi.org\/10.1109\/ISKE60036.2023.10480934","DOI":"10.1109\/ISKE60036.2023.10480934"},{"key":"971_CR24","doi-asserted-by":"crossref","unstructured":"Liu, Z., Huang, D., Huang, K., et\u00a0al. (2021). Finbert: A pre-trained financial language representation model for financial text mining. In: Proceedings of the twenty-ninth International Conference on International Joint Conferences on Artificial Intelligence, pp. 4513\u20134519, https:\/\/www.ijcai.org\/proceedings\/2020\/0622.pdf","DOI":"10.24963\/ijcai.2020\/622"},{"key":"971_CR25","doi-asserted-by":"publisher","unstructured":"Lu, W., Li, J., Li, Y., et\u00a0al. (2020). A cnn-lstm-based model to forecast stock prices. Complexity 2020(1), 6622927. https:\/\/doi.org\/10.1155\/2020\/6622927","DOI":"10.1155\/2020\/6622927"},{"key":"971_CR26","doi-asserted-by":"publisher","unstructured":"Lu, X., Qiu, J., Yang, Y., et\u00a0al. (2024). Large language model-based bidding behavior agent and market sentiment agent-assisted electricity price prediction. IEEE Transactions on Energy Markets, Policy and Regulation pp. 1\u201313. https:\/\/doi.org\/10.1109\/TEMPR.2024.3518624","DOI":"10.1109\/TEMPR.2024.3518624"},{"key":"971_CR27","doi-asserted-by":"publisher","unstructured":"Md, A. Q., Kapoor, S., A.V., C.J., et al. (2023). Novel optimization approach for stock price forecasting using multi-layered sequential lstm. Applied Soft Computing, 134, Article 109830. https:\/\/doi.org\/10.1016\/j.asoc.2022.109830","DOI":"10.1016\/j.asoc.2022.109830"},{"issue":"2","key":"971_CR28","doi-asserted-by":"publisher","first-page":"13","DOI":"10.5121\/ijcsea.2014.4202","volume":"4","author":"P Mondal","year":"2014","unstructured":"Mondal, P., Shit, L., & Goswami, S. (2014). Study of effectiveness of time series modeling (arima) in forecasting stock prices. International Journal of Computer Science, Engineering and Applications, 4(2), 13. https:\/\/doi.org\/10.5121\/ijcsea.2014.4202","journal-title":"International Journal of Computer Science, Engineering and Applications"},{"key":"971_CR29","doi-asserted-by":"publisher","unstructured":"Nie, Y., Nguyen, N.H., Sinthong, P., et\u00a0al. (2022). A time series is worth 64 words: Long-term forecasting with transformers. In: International Conference on Learning Representations, https:\/\/doi.org\/10.48550\/arXiv.2211.14730","DOI":"10.48550\/arXiv.2211.14730"},{"key":"971_CR30","doi-asserted-by":"publisher","unstructured":"Olorunnimbe, K., Viktor, H. (2024). Ensemble of temporal transformers for financial time series. Journal of Intelligent Information Systems 62(4), 1087\u20131111. https:\/\/doi.org\/10.1007\/s10844-024-00851-2","DOI":"10.1007\/s10844-024-00851-2"},{"key":"971_CR31","unstructured":"Ouyang, L., Wu, J., Jiang, X., et\u00a0al. (2022). Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems 35, 27730\u201327744. arXiv:2203.02155"},{"key":"971_CR32","unstructured":"Pi\u00f3ro, M., Ciebiera, K., Kr\u00f3l, K., et\u00a0al. (2024). Moe-mamba: Efficient selective state space models with mixture of experts. In: CoRR, arXiv:2401.04081"},{"key":"971_CR33","unstructured":"Smith, J.T.H., Warrington, A., Linderman, S.W. (2023). Simplified state space layers for sequence modeling. In: The Eleventh International Conference on Learning Representations, arXiv:2208.04933"},{"key":"971_CR34","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., et\u00a0al. (2017). Attention is all you need. In: Guyon, I., Luxburg, U.V., Bengio, S., et\u00a0al. (eds.) Advances in Neural Information Processing Systems, vol\u00a030. Curran Associates, Inc., https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2017\/file\/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf"},{"key":"971_CR35","doi-asserted-by":"publisher","unstructured":"Vo, Q.H., Nguyen, H.T., Le, B., et\u00a0al. (2017). Multi-channel lstm-cnn model for vietnamese sentiment analysis. In: 2017 9th International Conference on Knowledge and Systems Engineering, IEEE, pp. 24\u201329, https:\/\/doi.org\/10.1109\/KSE.2017.8119429","DOI":"10.1109\/KSE.2017.8119429"},{"key":"971_CR36","doi-asserted-by":"crossref","unstructured":"Wang, J., Yu, L.C., Lai, K.R., et\u00a0al. (2016). Dimensional sentiment analysis using a regional cnn-lstm model. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, pp. 225\u2013230, https:\/\/aclanthology.org\/P16-2037.pdf","DOI":"10.18653\/v1\/P16-2037"},{"key":"971_CR37","doi-asserted-by":"publisher","unstructured":"Wang, S., Wu, H., Shi, X., et\u00a0al. (2024a). Timemixer: Decomposable multiscale mixing for time series forecasting. https:\/\/doi.org\/10.48550\/arXiv.2405.14616","DOI":"10.48550\/arXiv.2405.14616"},{"key":"971_CR38","unstructured":"Wang, X., Feng, M., Qiu, J., et\u00a0al. (2024b). From news to forecast: Integrating event analysis in llm-based time series forecasting with reflection. In: Globerson, A., Mackey, L., Belgrave, D., et\u00a0al. (eds) Advances in Neural Information Processing Systems, 37. Curran Associates, Inc., pp. 58118\u201358153, https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2024\/file\/6aef8bffb372096ee73d98da30119f89-Paper-Conference.pdf"},{"key":"971_CR39","doi-asserted-by":"publisher","unstructured":"Wang, Z., Xie, Q., Feng, Y., et\u00a0al. (2023). Is chatgpt a good sentiment analyzer? a preliminary study. In: First Conference on Language Modeling, https:\/\/doi.org\/10.48550\/arXiv.2304.04339","DOI":"10.48550\/arXiv.2304.04339"},{"key":"971_CR40","doi-asserted-by":"publisher","unstructured":"Wang, Z., Kong, F., Feng, S., et al. (2025). Is mamba effective for time series forecasting? Neurocomputing, 619, Article 129178. https:\/\/doi.org\/10.1016\/j.neucom.2024.129178","DOI":"10.1016\/j.neucom.2024.129178"},{"key":"971_CR41","doi-asserted-by":"publisher","unstructured":"Yoo, J., Soun, Y., Park, Y.c., et\u00a0al. (2021). Accurate multivariate stock movement prediction via data-axis transformer with multi-level contexts. In: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining. Association for Computing Machinery, New York, NY, USA, KDD \u201921, p. 2037\u20132045, https:\/\/doi.org\/10.1145\/3447548.3467297","DOI":"10.1145\/3447548.3467297"},{"key":"971_CR42","doi-asserted-by":"publisher","unstructured":"Zeng, A., Chen, M., Zhang, L., et\u00a0al. (2023a). Are transformers effective for time series forecasting? In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 11121\u201311128, https:\/\/doi.org\/10.1609\/aaai.v37i9.26317","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"971_CR43","doi-asserted-by":"publisher","unstructured":"Zeng, L., Hu, H., Song, Q., et\u00a0al. (2024a). A drift-aware dynamic ensemble model with two-stage member selection for carbon price forecasting. Energy 313, 133699. https:\/\/doi.org\/10.1016\/j.energy.2024.133699","DOI":"10.1016\/j.energy.2024.133699"},{"key":"971_CR44","doi-asserted-by":"publisher","unstructured":"Zeng, L., Hu, H., Tang, H., et al. (2024). Carbon emission price point-interval forecasting based on multivariate variational mode decomposition and attention-lstm model. Applied Soft Computing, 157, Article 111543. https:\/\/doi.org\/10.1016\/j.asoc.2024.111543","DOI":"10.1016\/j.asoc.2024.111543"},{"key":"971_CR45","unstructured":"Zeng, Z., Kaur, R., Siddagangappa, S., et\u00a0al. (2023b). Financial time series forecasting using cnn and transformer. arXiv:2304.04912"},{"key":"971_CR46","unstructured":"Zhang, Y., Yan, J. (2023). Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In: The eleventh International Conference on Learning Representations, https:\/\/openreview.net\/forum?id=vSVLM2j9eie"},{"key":"971_CR47","doi-asserted-by":"publisher","unstructured":"Zheng, Y., Zhang, R., Zhang, J., et\u00a0al. (2024). Llamafactory: Unified efficient fine-tuning of 100+ language models. https:\/\/doi.org\/10.18653\/v1\/2024.acl-demos.38","DOI":"10.18653\/v1\/2024.acl-demos.38"},{"key":"971_CR48","unstructured":"Zhou, T., Ma, Z., Wen, Q., et\u00a0al. (2022). Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In: International Conference on Machine Learning, PMLR, pp. 27268\u201327286, https:\/\/proceedings.mlr.press\/v162\/zhou22g.html"}],"container-title":["Journal of Intelligent Information Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10844-025-00971-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10844-025-00971-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10844-025-00971-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,28]],"date-time":"2025-10-28T08:11:32Z","timestamp":1761639092000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10844-025-00971-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,2]]},"references-count":48,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["971"],"URL":"https:\/\/doi.org\/10.1007\/s10844-025-00971-3","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-6277319\/v1","asserted-by":"object"}]},"ISSN":["0925-9902","1573-7675"],"issn-type":[{"value":"0925-9902","type":"print"},{"value":"1573-7675","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,8,2]]},"assertion":[{"value":"21 March 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 July 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 July 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 August 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Approval"}}]}}