{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,16]],"date-time":"2025-09-16T17:33:25Z","timestamp":1758044005271,"version":"3.44.0"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2025,8,11]],"date-time":"2025-08-11T00:00:00Z","timestamp":1754870400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,8,11]],"date-time":"2025-08-11T00:00:00Z","timestamp":1754870400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Zhiqiang Zeng","award":["52305550"],"award-info":[{"award-number":["52305550"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Data Min Knowl Disc"],"published-print":{"date-parts":[[2025,9]]},"DOI":"10.1007\/s10618-025-01149-9","type":"journal-article","created":{"date-parts":[[2025,8,11]],"date-time":"2025-08-11T12:48:32Z","timestamp":1754916512000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Crossmamba: multivariate time series forecasting model for cross-temporal and cross-dimensional dependencies with Mamba"],"prefix":"10.1007","volume":"39","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-5194-2966","authenticated-orcid":false,"given":"Yuhan","family":"Lin","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7924-8655","authenticated-orcid":false,"given":"Liping","family":"Xiong","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5409-2962","authenticated-orcid":false,"given":"Zhiyong","family":"Hong","sequence":"additional","affiliation":[]},{"given":"Zhiqiang","family":"Zeng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0008-2731-4567","authenticated-orcid":false,"given":"Jian","family":"Zeng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0001-9722-6561","authenticated-orcid":false,"given":"Guoqiang","family":"Zeng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,11]]},"reference":[{"doi-asserted-by":"publisher","unstructured":"Ahamed MA, Cheng QS (2024) Timemachine: a time series is worth 4 mambas for long-term forecasting. In: Endriss U, Melo FS, Bach K, et\u00a0al (eds) ECAI 2024: 27th European conference on artificial intelligence, 19\u201324 October 2024, Santiago de Compostela, Spain\u2014including 13th conference on prestigious applications of intelligent systems (PAIS 2024), Frontiers in Artificial Intelligence and Applications, vol 392. IOS Press, pp 1688\u20131695. https:\/\/doi.org\/10.3233\/FAIA240677","key":"1149_CR1","DOI":"10.3233\/FAIA240677"},{"unstructured":"Das A, Kong W, Leach A et\u00a0al (2023) Long-term forecasting with tide: time-series dense encoder. Trans Mach Learn Res 2023","key":"1149_CR2"},{"issue":"5","key":"1149_CR3","first-page":"695","volume":"20","author":"\u00d6F Demirel","year":"2012","unstructured":"Demirel \u00d6F, Zaim S, \u00c7ali\u015fkan A et al (2012) Forecasting natural gas consumption in Istanbul using neural networks and multivariate time series methods. Turk J Electr Eng Comput Sci 20(5):695\u2013711","journal-title":"Turk J Electr Eng Comput Sci"},{"unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A et\u00a0al (2021) An image is worth 16x16 words: transformers for image recognition at scale. In: 9th international conference on learning representations, ICLR 2021, Virtual Event, Austria, May 3\u20137, 2021. OpenReview.net","key":"1149_CR4"},{"doi-asserted-by":"publisher","unstructured":"Du D, Su B, Wei Z (2023) Preformer: Predictive transformer with multi-scale segment-wise correlations for long-term time series forecasting. In: IEEE International Conference on Acoustics, Speech and Signal Processing ICASSP 2023, Rhodes Island, Greece, June 4-10, 2023. IEEE, pp 1\u20135, https:\/\/doi.org\/10.1109\/ICASSP49357.2023.10096881","key":"1149_CR5","DOI":"10.1109\/ICASSP49357.2023.10096881"},{"doi-asserted-by":"publisher","unstructured":"Dubey A, Jauhri A, Pandey A et\u00a0al (2024) The llama 3 herd of models. CoRR. https:\/\/doi.org\/10.48550\/ARXIV.2407.21783, arXiv:2407.21783","key":"1149_CR6","DOI":"10.48550\/ARXIV.2407.21783"},{"doi-asserted-by":"publisher","unstructured":"Gu A, Dao T (2023) Mamba: linear-time sequence modeling with selective state spaces. CoRR. https:\/\/doi.org\/10.48550\/ARXIV.2312.00752, arXiv:2312.00752","key":"1149_CR7","DOI":"10.48550\/ARXIV.2312.00752"},{"unstructured":"Ho J, Jain A, Abbeel P (2020) Denoising diffusion probabilistic models. In: Larochelle H, Ranzato M, Hadsell R et\u00a0al (eds) Advances in neural information processing systems 33: annual conference on neural information processing systems 2020, NeurIPS 2020, December 6\u201312, 2020, virtual","key":"1149_CR8"},{"doi-asserted-by":"publisher","unstructured":"Lea C, Flynn MD, Vidal R et\u00a0al (2017) Temporal convolutional networks for action segmentation and detection. In: 2017 IEEE conference on computer vision and pattern recognition, CVPR 2017, Honolulu, HI, USA, July 21\u201326, 2017. IEEE Computer Society, pp 1003\u20131012. https:\/\/doi.org\/10.1109\/CVPR.2017.113","key":"1149_CR9","DOI":"10.1109\/CVPR.2017.113"},{"doi-asserted-by":"publisher","unstructured":"Li L, Yan J, Yang X et\u00a0al (2019) Learning interpretable deep state space model for probabilistic time series forecasting. In: Kraus S (ed) Proceedings of the twenty-eighth international joint conference on artificial intelligence, IJCAI 2019, Macao, China, August 10\u201316, 2019. ijcai.org, pp 2901\u20132908. https:\/\/doi.org\/10.24963\/IJCAI.2019\/402","key":"1149_CR10","DOI":"10.24963\/IJCAI.2019\/402"},{"doi-asserted-by":"publisher","unstructured":"Li L, Zhang J, Yan J et\u00a0al (2021) Synergetic learning of heterogeneous temporal sequences for multi-horizon probabilistic forecasting. In: Thirty-fifth AAAI conference on artificial intelligence, AAAI 2021, thirty-third conference on innovative applications of artificial intelligence, IAAI 2021, the eleventh symposium on educational advances in artificial intelligence, EAAI 2021, Virtual Event, February 2\u20139, 2021. AAAI Press, pp 8420\u20138428. https:\/\/doi.org\/10.1609\/AAAI.V35I10.17023","key":"1149_CR11","DOI":"10.1609\/AAAI.V35I10.17023"},{"doi-asserted-by":"publisher","unstructured":"Li Z, Qi S, Li Y et\u00a0al (2023) Revisiting long-term time series forecasting: an investigation on linear mapping. CoRR. https:\/\/doi.org\/10.48550\/ARXIV.2305.10721, arXiv:2305.10721","key":"1149_CR12","DOI":"10.48550\/ARXIV.2305.10721"},{"unstructured":"Liu M, Zeng A, Chen M et\u00a0al (2022a) Scinet: time series modeling and forecasting with sample convolution and interaction. In: Koyejo S, Mohamed S, Agarwal A et\u00a0al (eds) Advances in neural information processing systems 35: annual conference on neural information processing systems 2022, NeurIPS 2022, New Orleans, LA, USA, November 28\u2013December 9, 2022","key":"1149_CR13"},{"unstructured":"Liu S, Yu H, Liao C et\u00a0al (2022b) Pyraformer: low-complexity pyramidal attention for long-range time series modeling and forecasting. In: The tenth international conference on learning representations, ICLR 2022, Virtual Event, April 25\u201329, 2022. OpenReview.net","key":"1149_CR14"},{"unstructured":"Liu Y, Wu H, Wang J et\u00a0al (2022c) Non-stationary transformers: exploring the stationarity in time series forecasting. In: Koyejo S, Mohamed S, Agarwal A, et\u00a0al (eds) Advances in neural information processing systems 35: annual conference on neural information processing systems 2022, NeurIPS 2022, New Orleans, LA, USA, November 28\u2013December 9, 2022","key":"1149_CR15"},{"unstructured":"Liu Y, Hu T, Zhang H et\u00a0al (2024) itransformer: inverted transformers are effective for time series forecasting. In: The twelfth international conference on learning representations, ICLR 2024, Vienna, Austria, May 7\u201311, 2024. OpenReview.net","key":"1149_CR16"},{"doi-asserted-by":"crossref","unstructured":"L\u00fctkepohl H (2013) Vector autoregressive models. In: Handbook of research methods and applications in empirical macroeconomics. Edward Elgar Publishing, pp 139\u2013164","key":"1149_CR17","DOI":"10.4337\/9780857931023.00012"},{"doi-asserted-by":"publisher","unstructured":"Narang S, Chung HW, Tay Y et\u00a0al (2021) Do transformer modifications transfer across implementations and applications? In: Moens M, Huang X, Specia L et\u00a0al (eds) Proceedings of the 2021 conference on empirical methods in natural language processing, EMNLP 2021, Virtual Event\/Punta Cana, Dominican Republic, 7\u201311 November, 2021. Association for Computational Linguistics, pp 5758\u20135773. https:\/\/doi.org\/10.18653\/V1\/2021.EMNLP-MAIN.465","key":"1149_CR18","DOI":"10.18653\/V1\/2021.EMNLP-MAIN.465"},{"unstructured":"Nie Y, Nguyen NH, Sinthong P et\u00a0al (2023) A time series is worth 64 words: long-term forecasting with transformers. In: The eleventh international conference on learning representations, ICLR 2023, Kigali, Rwanda, May 1\u20135 2023. OpenReview.net","key":"1149_CR19"},{"doi-asserted-by":"publisher","unstructured":"Patro BN, Agneeswaran VS (2024) Simba: simplified mamba-based architecture for vision and multivariate time series. CoRR. https:\/\/doi.org\/10.48550\/ARXIV.2403.15360, arXiv:2403.15360","key":"1149_CR20","DOI":"10.48550\/ARXIV.2403.15360"},{"key":"1149_CR21","doi-asserted-by":"publisher","first-page":"899","DOI":"10.1016\/B978-0-444-62731-5.00016-6","volume":"2","author":"A Patton","year":"2013","unstructured":"Patton A (2013) Copula methods for forecasting multivariate time series. Handb Econ Forecast 2:899\u2013960","journal-title":"Handb Econ Forecast"},{"doi-asserted-by":"publisher","unstructured":"Qin Y, Song D, Chen H et\u00a0al (2017) A dual-stage attention-based recurrent neural network for time series prediction. In: Sierra C (ed) Proceedings of the twenty-sixth international joint conference on artificial intelligence, IJCAI 2017, Melbourne, Australia, August 19\u201325 2017. ijcai.org, pp 2627\u20132633. https:\/\/doi.org\/10.24963\/IJCAI.2017\/366","key":"1149_CR22","DOI":"10.24963\/IJCAI.2017\/366"},{"key":"1149_CR23","first-page":"1401","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel C, Shazeer N, Roberts A et al (2020) Exploring the limits of transfer learning with a unified text-to-text transformer. J Mach Learn Res 21:1401\u201314067","journal-title":"J Mach Learn Res"},{"unstructured":"Rangapuram SS, Seeger MW, Gasthaus J et\u00a0al (2018) Deep state space models for time series forecasting. In: Bengio S, Wallach HM, Larochelle H et\u00a0al (eds) Advances in neural information processing systems 31: annual conference on neural information processing systems 2018, NeurIPS 2018, December 3\u20138 2018, Montr\u00e9al, Canada, pp 7796\u20137805","key":"1149_CR24"},{"issue":"6","key":"1149_CR25","doi-asserted-by":"publisher","first-page":"386","DOI":"10.1037\/h0042519","volume":"65","author":"F Rosenblatt","year":"1958","unstructured":"Rosenblatt F (1958) The perceptron: a probabilistic model for information storage and organization in the brain. Psychol Rev 65(6):386","journal-title":"Psychol Rev"},{"issue":"3","key":"1149_CR26","doi-asserted-by":"publisher","first-page":"1181","DOI":"10.1016\/j.ijforecast.2019.07.001","volume":"36","author":"D Salinas","year":"2020","unstructured":"Salinas D, Flunkert V, Gasthaus J et al (2020) Deepar: probabilistic forecasting with autoregressive recurrent networks. Int J Forecast 36(3):1181\u20131191","journal-title":"Int J Forecast"},{"doi-asserted-by":"crossref","unstructured":"Scherrer W, Deistler M (2019) Vector autoregressive moving average models. In: Handbook of statistics, vol\u00a041. Elsevier, pp 145\u2013191","key":"1149_CR27","DOI":"10.1016\/bs.host.2019.01.004"},{"doi-asserted-by":"publisher","unstructured":"Touvron H, Lavril T, Izacard G et\u00a0al (2023a) Llama: open and efficient foundation language models. CoRR. https:\/\/doi.org\/10.48550\/ARXIV.2302.13971, arXiv:2302.13971","key":"1149_CR28","DOI":"10.48550\/ARXIV.2302.13971"},{"doi-asserted-by":"publisher","unstructured":"Touvron H, Martin L, Stone K et\u00a0al (2023b) Llama 2: open foundation and fine-tuned chat models. CoRR. https:\/\/doi.org\/10.48550\/ARXIV.2307.09288, arXiv:2307.09288","key":"1149_CR29","DOI":"10.48550\/ARXIV.2307.09288"},{"unstructured":"Vaswani A, Shazeer N, Parmar N et\u00a0al (2017) Attention is all you need. In: Guyon I, von Luxburg U, Bengio S et\u00a0al (eds) Advances in neural information processing systems 30: annual conference on neural information processing systems 2017, December 4\u20139, 2017, Long Beach, CA, USA, pp 5998\u20136008","key":"1149_CR30"},{"key":"1149_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/J.NEUCOM.2024.129178","volume":"619","author":"Z Wang","year":"2025","unstructured":"Wang Z, Kong F, Feng S et al (2025) Is mamba effective for time series forecasting? Neurocomputing 619:129178. https:\/\/doi.org\/10.1016\/J.NEUCOM.2024.129178","journal-title":"Neurocomputing"},{"unstructured":"Wu H, Xu J, Wang J et\u00a0al (2021) Autoformer: decomposition transformers with auto-correlation for long-term series forecasting. In: Ranzato M, Beygelzimer A, Dauphin YN et\u00a0al (eds) Advances in neural information processing systems 34: annual conference on neural information processing systems 2021, NeurIPS 2021, December 6\u201314, 2021, virtual, pp 22419\u201322430","key":"1149_CR32"},{"unstructured":"Wu H, Hu T, Liu Y et\u00a0al (2023) Timesnet: temporal 2d-variation modeling for general time series analysis. In: The eleventh international conference on learning representations, ICLR 2023, Kigali, Rwanda, May 1\u20135, 2023. OpenReview.net","key":"1149_CR33"},{"doi-asserted-by":"publisher","unstructured":"Wu Z, Pan S, Long G et\u00a0al (2020) Connecting the dots: multivariate time series forecasting with graph neural networks. In: Gupta R, Liu Y, Tang J et\u00a0al (eds) KDD \u201920: the 26th ACM SIGKDD conference on knowledge discovery and data mining, virtual event, CA, USA, August 23\u201327, 2020. ACM, pp 753\u2013763. https:\/\/doi.org\/10.1145\/3394486.3403118","key":"1149_CR34","DOI":"10.1145\/3394486.3403118"},{"key":"1149_CR35","first-page":"69638","volume":"36","author":"K Yi","year":"2023","unstructured":"Yi K, Zhang Q, Fan W et al (2023) Fouriergnn: rethinking multivariate time series forecasting from a pure graph perspective. Adv Neural Inf Process Syst 36:69638\u201369660","journal-title":"Adv Neural Inf Process Syst"},{"key":"1149_CR36","first-page":"76656","volume":"36","author":"K Yi","year":"2023","unstructured":"Yi K, Zhang Q, Fan W et al (2023) Frequency-domain mlps are more effective learners in time series forecasting. Adv Neural Inf Process Syst 36:76656\u201376679","journal-title":"Adv Neural Inf Process Syst"},{"key":"1149_CR37","first-page":"55115","volume":"37","author":"K Yi","year":"2024","unstructured":"Yi K, Fei J, Zhang Q et al (2024) Filternet: harnessing frequency filters for time series forecasting. Adv Neural Inf Process Syst 37:55115\u201355140","journal-title":"Adv Neural Inf Process Syst"},{"doi-asserted-by":"publisher","unstructured":"Zeng A, Chen M, Zhang L et\u00a0al (2023) Are transformers effective for time series forecasting? In: Williams B, Chen Y, Neville J (eds) Thirty-seventh aaai conference on artificial intelligence, AAAI 2023, thirty-fifth conference on innovative applications of artificial intelligence, IAAI 2023, Thirteenth symposium on educational advances in artificial intelligence, EAAI 2023, Washington, DC, USA, February 7\u201314, 2023. AAAI Press, pp 11121\u201311128. https:\/\/doi.org\/10.1609\/AAAI.V37I9.26317","key":"1149_CR38","DOI":"10.1609\/AAAI.V37I9.26317"},{"unstructured":"Zhang B, Sennrich R (2019) Root mean square layer normalization. In: Wallach HM, Larochelle H, Beygelzimer A et\u00a0al (eds) Advances in neural information processing systems 32: annual conference on neural information processing systems 2019, NeurIPS 2019, December 8\u201314, 2019, Vancouver, BC, Canada, pp 12360\u201312371","key":"1149_CR39"},{"unstructured":"Zhang Y, Yan J (2023) Crossformer: transformer utilizing cross-dimension dependency for multivariate time series forecasting. In: The eleventh international conference on learning representations, ICLR 2023, Kigali, Rwanda, May 1\u20135, 2023. OpenReview.net","key":"1149_CR40"},{"doi-asserted-by":"publisher","unstructured":"Zhou H, Zhang S, Peng J et\u00a0al (2021) Informer: beyond efficient transformer for long sequence time-series forecasting. In: Thirty-fifth AAAI conference on artificial intelligence, AAAI 2021, thirty-third conference on innovative applications of artificial intelligence, IAAI 2021, the eleventh symposium on educational advances in artificial intelligence, EAAI 2021, Virtual Event, February 2\u20139, 2021. AAAI Press, pp 11106\u201311115. https:\/\/doi.org\/10.1609\/AAAI.V35I12.17325","key":"1149_CR41","DOI":"10.1609\/AAAI.V35I12.17325"},{"unstructured":"Zhou T, Ma Z, Wen Q et\u00a0al (2022) Fedformer: frequency enhanced decomposed transformer for long-term series forecasting. In: Chaudhuri K, Jegelka S, Song L et\u00a0al (eds) International conference on machine learning, ICML 2022, 17\u201323 July 2022, Baltimore, Maryland, USA, Proceedings of machine learning research, vol 162. PMLR, pp 27268\u201327286","key":"1149_CR42"}],"container-title":["Data Mining and Knowledge Discovery"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10618-025-01149-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10618-025-01149-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10618-025-01149-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,12]],"date-time":"2025-09-12T10:30:13Z","timestamp":1757673013000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10618-025-01149-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,11]]},"references-count":42,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,9]]}},"alternative-id":["1149"],"URL":"https:\/\/doi.org\/10.1007\/s10618-025-01149-9","relation":{},"ISSN":["1384-5810","1573-756X"],"issn-type":[{"type":"print","value":"1384-5810"},{"type":"electronic","value":"1573-756X"}],"subject":[],"published":{"date-parts":[[2025,8,11]]},"assertion":[{"value":"9 May 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 July 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 August 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"68"}}