{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T10:11:19Z","timestamp":1764843079522,"version":"3.46.0"},"reference-count":51,"publisher":"Springer Science and Business Media LLC","issue":"18","license":[{"start":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T00:00:00Z","timestamp":1764806400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T00:00:00Z","timestamp":1764806400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Supercomput"],"DOI":"10.1007\/s11227-025-08115-8","type":"journal-article","created":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T10:05:19Z","timestamp":1764842719000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["BiSTAG-TS: a dual-stream generative framework for symbolic\u2013numerical time series forecasting via large language models"],"prefix":"10.1007","volume":"81","author":[{"given":"Ruidi","family":"Yang","sequence":"first","affiliation":[]},{"given":"Yuxing","family":"Mao","sequence":"additional","affiliation":[]},{"given":"Hengyu","family":"Yan","sequence":"additional","affiliation":[]},{"given":"Zijie","family":"Wei","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jianyu","family":"Pan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,4]]},"reference":[{"key":"8115_CR1","unstructured":"Shumway RH, Stoffer DS, Shumway RH, Stoffer DS (2017) Arima models. Time series analysis and its applications: with R examples 75\u2013163"},{"key":"8115_CR2","doi-asserted-by":"publisher","DOI":"10.2307\/j.ctv14jx6sm","volume-title":"Time Series Analysis","author":"JD Hamilton","year":"2020","unstructured":"Hamilton JD (2020) Time Series Analysis. Princeton University Press, Princeton"},{"issue":"1","key":"8115_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1162\/neco.1997.9.1.1","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter S, Schmidhuber J (1997) Flat minima. Neural Comput 9(1):1\u201342","journal-title":"Neural Comput"},{"key":"8115_CR4","unstructured":"Borovykh A, Bohte S, Oosterlee CW (2017) Conditional time series forecasting with convolutional neural networks. arXiv preprint arXiv:1703.04691"},{"key":"8115_CR5","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser \u0141, Polosukhin I (2017) Attention is all you need. Advances in neural information processing systems 30"},{"key":"8115_CR6","unstructured":"Zhou H, Zhang S, Peng J, Zhang S, Li J, Xiong H, Zhang W (2021) Informer: Beyond efficient transformer for long sequence time-series forecasting. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 35, pp 11106\u201311115"},{"key":"8115_CR7","first-page":"22419","volume":"34","author":"H Wu","year":"2021","unstructured":"Wu H, Xu J, Wang J, Long M (2021) Autoformer: decomposition transformers with auto-correlation for long-term series forecasting. Adv Neural Inf Process Syst 34:22419\u201322430","journal-title":"Adv Neural Inf Process Syst"},{"key":"8115_CR8","unstructured":"Nie Y, Nguyen NH, Sinthong P, Kalagnanam J (2022) A time series is worth 64 words: Long-term forecasting with transformers. arXiv preprint arXiv:2211.14730"},{"key":"8115_CR9","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A et al (2020) Language models are few-shot learners. Adv Neural Inf Process Syst 33:1877\u20131901","journal-title":"Adv Neural Inf Process Syst"},{"issue":"1","key":"8115_CR10","doi-asserted-by":"publisher","first-page":"254","DOI":"10.31181\/jscda11202324","volume":"1","author":"A \u00d6nden","year":"2023","unstructured":"\u00d6nden A, Alnour M (2023) ChatGPT and OpenAI: a comprehensive bibliometric review. J Soft Comput Decis Anal 1(1):254\u2013264","journal-title":"J Soft Comput Decis Anal"},{"key":"8115_CR11","unstructured":"Zhang X, Chowdhury RR, Gupta RK, Shang J (2024) Large language models for time series: A survey. arXiv preprint arXiv:2402.01801"},{"key":"8115_CR12","unstructured":"Jin M, Wang S, Ma L, Chu Z, Zhang JY, Shi X, Chen P-Y, Liang Y, Li Y-F, Pan S et al (2023) Time-llm: Time series forecasting by reprogramming large language models. arXiv preprint arXiv:2310.01728"},{"key":"8115_CR13","unstructured":"Liu C, Xu Q, Miao H, Yang S, Zhang L, Long C, Li Z, Zhao R (2025) Timecma: Towards llm-empowered multivariate time series forecasting via cross-modality alignment. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 39, pp 18780\u201318788"},{"key":"8115_CR14","unstructured":"Zhang X, Feng S, Li X (2025) From text to time? Rethinking the effectiveness of the large language model for time series forecasting. arXiv preprint arXiv:2504.08818"},{"key":"8115_CR15","unstructured":"Ye J, Zhang W, Yi K, Yu Y, Li Z, Li J, Tsung F (2024) A survey of time series foundation models: Generalizing time series representation with large language model. arXiv preprint arXiv:2405.02358"},{"key":"8115_CR16","unstructured":"Chen C, Oliveira G, Noghabi HS, Sylvain T (2024) Llm-ts integrator: Integrating llm for enhanced time series modeling. arXiv preprint arXiv:2410.16489"},{"key":"8115_CR17","unstructured":"Wang C, Qi Q, Wang J, Sun H, Zhuang Z, Wu J, Zhang L, Liao J (2025) Chattime: A unified multimodal time series foundation model bridging numerical and textual data. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 39, pp 12694\u201312702"},{"key":"8115_CR18","unstructured":"Chen J, Feng A, Zhao Z, Garza J, Nurbek G, Qin C, Maatouk A, Tassiulas L, Gao Y, Ying R (2025) Mtbench: A multimodal time series benchmark for temporal reasoning and question answering. arXiv preprint arXiv:2503.16858"},{"key":"8115_CR19","unstructured":"Kong Y, Yang Y, Hwang Y, Du W, Zohren S, Wang Z, Jin M, Wen Q (2025) Time-mqa: Time series multi-task question answering with context enhancement. arXiv preprint arXiv:2503.01875"},{"key":"8115_CR20","unstructured":"Kong Y, Yang Y, Wang S, Liu C, Liang Y, Jin M, Zohren S, Pei D, Liu Y, Wen Q (2025) Position: Empowering time series reasoning with multimodal llms. arXiv preprint arXiv:2502.01477"},{"key":"8115_CR21","first-page":"122154","volume":"37","author":"Y Liu","year":"2024","unstructured":"Liu Y, Qin G, Huang X, Wang J, Long M (2024) Autotimes: autoregressive time series forecasters via large language models. Adv Neural Inf Process Syst 37:122154\u2013122184","journal-title":"Adv Neural Inf Process Syst"},{"key":"8115_CR22","unstructured":"Liu X, Hu J, Li Y, Diao S, Liang Y, Hooi B, Zimmermann R (2024) Unitime: A language-empowered unified model for cross-domain time series forecasting. In: Proceedings of the ACM Web Conference 2024, pp 4095\u20134106"},{"key":"8115_CR23","unstructured":"Xiao M, Jiang Z, Qian L, Chen Z, He Y, Xu Y, Jiang Y, Li D, Weng R-L, Peng M et al (2025) Enhancing financial time-series forecasting with retrieval-augmented large language models. arXiv preprint arXiv:2503.67890"},{"issue":"4","key":"8115_CR24","doi-asserted-by":"publisher","first-page":"637","DOI":"10.1016\/j.ijforecast.2006.03.005","volume":"22","author":"ES Gardner Jr","year":"2006","unstructured":"Gardner ES Jr (2006) Exponential smoothing: the state of the art-part II. Int J Forecast 22(4):637\u2013666","journal-title":"Int J Forecast"},{"key":"8115_CR25","doi-asserted-by":"publisher","DOI":"10.1093\/acprof:oso\/9780199641178.001.0001","volume-title":"Time Series Analysis by State Space Methods","author":"J Durbin","year":"2012","unstructured":"Durbin J, Koopman SJ (2012) Time Series Analysis by State Space Methods. Oxford University Press, Oxford"},{"key":"8115_CR26","unstructured":"Zeng A, Chen M, Zhang L, Xu Q (2023) Are transformers effective for time series forecasting? In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 37, pp 11121\u201311128"},{"issue":"2194","key":"8115_CR27","doi-asserted-by":"publisher","first-page":"20200209","DOI":"10.1098\/rsta.2020.0209","volume":"379","author":"B Lim","year":"2021","unstructured":"Lim B, Zohren S (2021) Time-series forecasting with deep learning: a survey. Phil Trans R Soc A 379(2194):20200209","journal-title":"Phil Trans R Soc A"},{"issue":"1","key":"8115_CR28","doi-asserted-by":"publisher","first-page":"388","DOI":"10.1016\/j.ijforecast.2020.06.008","volume":"37","author":"H Hewamalage","year":"2021","unstructured":"Hewamalage H, Bergmeir C, Bandara K (2021) Recurrent neural networks for time series forecasting: Current status and future directions. Int J Forecast 37(1):388\u2013427","journal-title":"Int J Forecast"},{"key":"8115_CR29","unstructured":"Wu H, Hu T, Liu Y, Zhou H, Wang J, Long M (2022) Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv preprint arXiv:2210.02186"},{"key":"8115_CR30","unstructured":"Liu Y, Hu T, Zhang H, Wu H, Wang S, Ma L, Long M (2023) itransformer: Inverted transformers are effective for time series forecasting. arXiv preprint arXiv:2310.06625"},{"key":"8115_CR31","unstructured":"Ekambaram V, Jati A, Nguyen N, Sinthong P, Kalagnanam J (2023) Tsmixer: Lightweight mlp-mixer model for multivariate time series forecasting. In: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp 459\u2013469"},{"key":"8115_CR32","unstructured":"Luo D, Cheng W, Wang Y, Xu D, Ni J, Yu W, Zhang X, Liu Y, Chen Y, Chen H et al (2023) Time series contrastive learning with information-aware augmentations. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 37, pp 4534\u20134542"},{"key":"8115_CR33","unstructured":"Kim T, Kim J, Tae Y, Park C, Choi J-H, Choo J (2021) Reversible instance normalization for accurate time-series forecasting against distribution shift. In: International Conference on Learning Representations"},{"key":"8115_CR34","unstructured":"Lu K, Grover A, Abbeel P, Mordatch I (2022) Frozen pretrained transformers as universal computation engines. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 36, pp 7628\u20137636"},{"issue":"11","key":"8115_CR35","doi-asserted-by":"publisher","first-page":"6851","DOI":"10.1109\/TKDE.2023.3342137","volume":"36","author":"H Xue","year":"2023","unstructured":"Xue H, Salim FD (2023) Promptcast: a new prompt-based learning paradigm for time series forecasting. IEEE Trans Knowl Data Eng 36(11):6851\u20136864","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"8115_CR36","first-page":"2403","volume":"36","author":"S Dooley","year":"2023","unstructured":"Dooley S, Khurana GS, Mohapatra C, Naidu SV, White C (2023) Forecastpfn: synthetically-trained zero-shot forecasting. Adv Neural Inf Process Syst 36:2403\u20132426","journal-title":"Adv Neural Inf Process Syst"},{"key":"8115_CR37","unstructured":"Sun C, Li H, Li Y, Hong S (2023) Test: Text prototype aligned embedding to activate llm\u2019s ability for time series. arXiv preprint arXiv:2308.08241"},{"key":"8115_CR38","unstructured":"Chang C, Peng W-C, Chen T-F (2023) Llm4ts: Two-stage fine-tuning for time-series forecasting with pre-trained llms. CoRR"},{"key":"8115_CR39","unstructured":"Ansari AF, Stella L, Turkmen C, Zhang X, Mercado P, Shen H, Shchur O, Rangapuram SS, Arango SP, Kapoor S et al (2024) Chronos: Learning the language of time series. arXiv preprint arXiv:2403.07815"},{"key":"8115_CR40","unstructured":"Rasul K, Ashok A, Williams AR, Khorasani A, Adamopoulos G, Bhagwatkar R, Bilo\u0161 M, Ghonia H, Hassen N, Schneider A et al (2023) Lag-llama: Towards foundation models for time series forecasting. In: R0-FoMo: Robustness of Few-shot and Zero-shot Learning in Large Foundation Models"},{"key":"8115_CR41","unstructured":"Garza A, Challu C, Mergenthaler-Canseco M (2023) Timegpt-1. arXiv preprint arXiv:2310.03589"},{"key":"8115_CR42","unstructured":"Liu Y, Zhang H, Li C, Huang X, Wang J, Long M (2024) Timer: Generative pre-trained transformers are large time series models. arXiv preprint arXiv:2402.02368"},{"key":"8115_CR43","unstructured":"Xie Z, Li Z, He X, Xu L, Wen X, Zhang T, Chen J, Shi R, Pei D (2024) Chatts: Aligning time series with llms via synthetic data for enhanced understanding and reasoning. arXiv preprint arXiv:2412.03104"},{"key":"8115_CR44","unstructured":"Zhang H, Yang C, Han J, Qin L, Wang X (2025) Tempogpt: Enhancing temporal reasoning via quantizing embedding. arXiv preprint arXiv:2501.07335"},{"key":"8115_CR45","unstructured":"Cai Y, Choudhry A, Goswami M, Dubrawski A (2024) Timeseriesexam: A time series understanding exam. arXiv preprint arXiv:2410.14752"},{"key":"8115_CR46","unstructured":"Liu Z, Wang T, Shi J, Zheng X, Chen Z, Song L, Dong W, Obeysekera J, Shirani F, Luo D (2024) Timex++: Learning time-series explanations with information bottleneck. arXiv preprint arXiv:2405.09308"},{"key":"8115_CR47","unstructured":"Jiang Y, Yu W, Lee G, Song D, Shin K, Cheng W, Liu Y, Chen H (2025) Explainable multi-modal time series prediction with llm-in-the-loop. arXiv preprint arXiv:2503.01013"},{"key":"8115_CR48","unstructured":"Godahewa R, Bergmeir C, Webb GI, Hyndman RJ, Montero-Manso P (2021) Monash time series forecasting archive. arXiv preprint arXiv:2105.06643"},{"key":"8115_CR49","unstructured":"Qiu X, Hu J, Zhou L, Wu X, Du J, Zhang B, Guo C, Zhou A, Jensen CS, Sheng Z et al (2024) Tfb: Towards comprehensive and fair benchmarking of time series forecasting methods. arXiv preprint arXiv:2403.20150"},{"key":"8115_CR50","first-page":"2825","volume":"12","author":"F Pedregosa","year":"2011","unstructured":"Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, Blondel M, Prettenhofer P, Weiss R, Dubourg V et al (2011) Scikit-learn: machine learning in python. J Mach Learn Res 12:2825\u20132830","journal-title":"J Mach Learn Res"},{"key":"8115_CR51","unstructured":"Taori R, Gulrajani I, Zhang T, Dubois Y, Li X, Guestrin C, Liang P, Alpaca THS (2023) An Instruction-Following LLaMA Model"}],"container-title":["The Journal of Supercomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-025-08115-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11227-025-08115-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-025-08115-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,4]],"date-time":"2025-12-04T10:05:20Z","timestamp":1764842720000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11227-025-08115-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,4]]},"references-count":51,"journal-issue":{"issue":"18","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["8115"],"URL":"https:\/\/doi.org\/10.1007\/s11227-025-08115-8","relation":{},"ISSN":["1573-0484"],"issn-type":[{"value":"1573-0484","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,4]]},"assertion":[{"value":"13 August 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 November 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"Not applicable.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Code availability"}}],"article-number":"1628"}}