{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T09:46:40Z","timestamp":1765619200765,"version":"3.48.0"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2025,2,7]],"date-time":"2025-02-07T00:00:00Z","timestamp":1738886400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,2,7]],"date-time":"2025-02-07T00:00:00Z","timestamp":1738886400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"The National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["No.62376081"],"award-info":[{"award-number":["No.62376081"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"The National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["No.U21A20484"],"award-info":[{"award-number":["No.U21A20484"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. Mach. Learn. &amp; Cyber."],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s13042-025-02530-2","type":"journal-article","created":{"date-parts":[[2025,2,7]],"date-time":"2025-02-07T02:14:01Z","timestamp":1738894441000},"page":"10081-10095","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A channel dependency decoupled two-stream model for multivariate time series analysis"],"prefix":"10.1007","volume":"16","author":[{"given":"Jin","family":"Fan","sequence":"first","affiliation":[]},{"given":"Jie","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Huifeng","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Jiaqian","family":"Xiang","sequence":"additional","affiliation":[]},{"given":"Yuxia","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Guanhang","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Jia","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,7]]},"reference":[{"key":"2530_CR1","first-page":"22419","volume":"34","author":"H Wu","year":"2021","unstructured":"Wu H, Xu J, Wang J, Long M (2021) Autoformer: decomposition transformers with auto-correlation for long-term series forecasting. Adv Neural Inf Process Syst 34:22419\u201322430","journal-title":"Adv Neural Inf Process Syst"},{"key":"2530_CR2","unstructured":"Wu H, Hu T, Liu Y, Zhou H, Wang J, Long M (2023) Timesnet: Temporal 2d-variation modeling for general time series analysis. International Conference on Learning Representations"},{"key":"2530_CR3","unstructured":"Xu J, Wu H, Wang J, Long M (2021) Anomaly transformer: Time series anomaly detection with association discrepancy. In: International Conference on Learning Representations"},{"key":"2530_CR4","doi-asserted-by":"publisher","first-page":"899","DOI":"10.1016\/B978-0-444-62731-5.00016-6","volume":"2","author":"A Patton","year":"2013","unstructured":"Patton A (2013) Copula methods for forecasting multivariate time series. Handbook Econ Forecast 2:899\u2013960","journal-title":"Handbook Econ Forecast"},{"key":"2530_CR5","doi-asserted-by":"crossref","unstructured":"Shao Z, Zhang Z, Wang F, Xu Y (2022) Pre-training enhanced spatial-temporal graph neural network for multivariate time series forecasting. In: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 1567\u20131577","DOI":"10.1145\/3534678.3539396"},{"key":"2530_CR6","unstructured":"Liu Y, Hu T, Zhang H, Wu H, Wang S, Ma L, Long M (2023) itransformer: Inverted transformers are effective for time series forecasting. arXiv preprint arXiv:2310.06625"},{"issue":"3","key":"2530_CR7","doi-asserted-by":"publisher","first-page":"1181","DOI":"10.1016\/j.ijforecast.2019.07.001","volume":"36","author":"D Salinas","year":"2020","unstructured":"Salinas D, Flunkert V, Gasthaus J, Januschowski T (2020) Deepar: Probabilistic forecasting with autoregressive recurrent networks. Int J Forecast 36(3):1181\u20131191","journal-title":"Int J Forecast"},{"key":"2530_CR8","doi-asserted-by":"crossref","unstructured":"Lim B, Zohren S, Roberts S (2020) Recurrent neural filters: Learning independent bayesian filtering steps for time series prediction. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1\u20138 . IEEE","DOI":"10.1109\/IJCNN48605.2020.9206906"},{"key":"2530_CR9","unstructured":"Oord A, Dieleman S, Zen H, Simonyan K, Vinyals O, Graves A, Kalchbrenner N, Senior A, Kavukcuoglu K Wavenet: A generative model for raw audio. In: 9th ISCA Speech Synthesis Workshop, pp. 125\u2013125"},{"key":"2530_CR10","unstructured":"Bai S, Kolter J Z, Koltun V (2018) An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv:1803.01271"},{"key":"2530_CR11","first-page":"11106","volume":"35","author":"H Zhou","year":"2021","unstructured":"Zhou H, Zhang S, Peng J, Zhang S, Li J, Xiong H, Zhang W (2021) Informer: beyond efficient transformer for long sequence time-series forecasting. Proceed AAAI Conf Artif Intell 35:11106\u201311115","journal-title":"Proceed AAAI Conf Artif Intell"},{"key":"2530_CR12","unstructured":"Zhou T, Ma Z, Wen Q, Wang X, Sun L, Jin R (2022) Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In: International Conference on Machine Learning, pp. 27268\u201327286 . PMLR"},{"key":"2530_CR13","unstructured":"Woo G, Liu C, Sahoo D, Kumar A, Hoi S (2022) Etsformer: Exponential smoothing transformers for time-series forecasting. arXiv preprint arXiv:2202.01381"},{"issue":"3","key":"2530_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3469087","volume":"13","author":"S Wang","year":"2022","unstructured":"Wang S, Zhang M, Miao H, Peng Z, Yu PS (2022) Multivariate correlation-aware spatio-temporal graph convolutional networks for multi-scale traffic prediction. ACM Trans Intell Syst Technol (TIST) 13(3):1\u201322","journal-title":"ACM Trans Intell Syst Technol (TIST)"},{"key":"2530_CR15","doi-asserted-by":"crossref","unstructured":"Du J, Wang S, Miao H, Zhang J (2021) Multi-channel pooling graph neural networks. In: IJCAI, pp. 1442\u20131448","DOI":"10.24963\/ijcai.2021\/199"},{"key":"2530_CR16","first-page":"11121","volume":"37","author":"A Zeng","year":"2023","unstructured":"Zeng A, Chen M, Zhang L, Xu Q (2023) Are transformers effective for time series forecasting? Proceed AAAI Conf Artif Intell 37:11121\u201311128","journal-title":"Proceed AAAI Conf Artif Intell"},{"key":"2530_CR17","unstructured":"Zhang Y, Yan J (2023) Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In: The Eleventh International Conference on Learning Representations"},{"key":"2530_CR18","doi-asserted-by":"crossref","unstructured":"Zhou P, Han X, Morariu V I, Davis L S (2017) Two-stream neural networks for tampered face detection. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1831\u20131839. IEEE","DOI":"10.1109\/CVPRW.2017.229"},{"key":"2530_CR19","doi-asserted-by":"crossref","unstructured":"Wang H, Wang L (2017) Modeling temporal dynamics and spatial configurations of actions using two-stream recurrent neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 499\u2013508","DOI":"10.1109\/CVPR.2017.387"},{"key":"2530_CR20","unstructured":"Wang L, Xiong Y, Wang Z, Qiao Y (2015) Towards good practices for very deep two-stream convnets. arXiv preprint arXiv:1507.02159"},{"key":"2530_CR21","doi-asserted-by":"crossref","unstructured":"Feichtenhofer C, Fan H, Malik J, He K (2019) Slowfast networks for video recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6202\u20136211","DOI":"10.1109\/ICCV.2019.00630"},{"key":"2530_CR22","unstructured":"Simonyan K, Zisserman A (2014) Two-stream convolutional networks for action recognition in videos. Adv Neural Inform Process Syst 27"},{"key":"2530_CR23","doi-asserted-by":"crossref","unstructured":"Guo H, Tang R, Ye Y, Li Z, He X (2017) Deepfm: a factorization-machine based neural network for ctr prediction. arXiv preprint arXiv:1703.04247","DOI":"10.24963\/ijcai.2017\/239"},{"key":"2530_CR24","doi-asserted-by":"crossref","unstructured":"Song W, Shi C, Xiao Z, Duan Z, Xu Y, Zhang M, Tang J (2019) Autoint: Automatic feature interaction learning via self-attentive neural networks. In: Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 1161\u20131170","DOI":"10.1145\/3357384.3357925"},{"key":"2530_CR25","doi-asserted-by":"crossref","unstructured":"Mao K, Zhu J, Su L, Cai G, Li Y, Dong Z (2023) Finalmlp: An enhanced two-stream mlp model for ctr prediction. arXiv preprint arXiv:2304.00902","DOI":"10.1609\/aaai.v37i4.25577"},{"key":"2530_CR26","doi-asserted-by":"crossref","unstructured":"Rendle S (2010) Factorization machines. In: 2010 IEEE International Conference on Data Mining, pp. 995\u20131000. IEEE","DOI":"10.1109\/ICDM.2010.127"},{"key":"2530_CR27","doi-asserted-by":"crossref","unstructured":"Wang R, Fu B, Fu G, Wang M (2017) Deep & cross network for ad click predictions. In: Proceedings of the ADKDD\u201917, pp. 1\u20137","DOI":"10.1145\/3124749.3124754"},{"key":"2530_CR28","doi-asserted-by":"crossref","unstructured":"Yu F, Liu Z, Liu Q, Zhang H, Wu S, Wang, L (2020) Deep interaction machine: A simple but effective model for high-order feature interactions. In: Proceedings of the 29th ACM International Conference on Information & Knowledge Management, pp. 2285\u20132288","DOI":"10.1145\/3340531.3412077"},{"key":"2530_CR29","unstructured":"Oreshkin B N, Carpov D, Chapados N, Bengio Y (2019) N-beats: Neural basis expansion analysis for interpretable time series forecasting. In: International Conference on Learning Representations"},{"key":"2530_CR30","doi-asserted-by":"crossref","unstructured":"Challu C, Olivares K G, Oreshkin B N, Ramirez F G, Canseco M M, Dubrawski A (2023)Nhits: Neural hierarchical interpolation for time series forecasting. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, pp. 6989\u20136997","DOI":"10.1609\/aaai.v37i6.25854"},{"key":"2530_CR31","unstructured":"Franceschi J -Y, Dieuleveut A, Jaggi M (2019) Unsupervised scalable representation learning for multivariate time series. Adv Neural Inform Process Syst 32"},{"key":"2530_CR32","doi-asserted-by":"crossref","unstructured":"Lai G, Chang W -C, Yang Y, Liu H (2018) Modeling long-and short-term temporal patterns with deep neural networks. In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 95\u2013104","DOI":"10.1145\/3209978.3210006"},{"key":"2530_CR33","unstructured":"Choromanski K M, Likhosherstov V, Dohan D, Song X, Gane A, Sarlos T, Hawkins P, Davis J Q, Mohiuddin A, Kaiser L (2020) Rethinking attention with performers. In: International Conference on Learning Representations"},{"key":"2530_CR34","unstructured":"Li S, Jin X, Xuan Y, Zhou X, Chen W, Wang Y -X, Yan X (2019) Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Adv Neural Inform Process Syst 32"},{"key":"2530_CR35","unstructured":"Zhang X, Jin X, Gopalswamy K, Gupta G, Park Y, Shi X, Wang H, Maddix D C, Wang B (2022) First de-trend then attend: Rethinking attention for time-series forecasting. In: NeurIPS\u201922 Workshop on All Things Attention: Bridging Different Perspectives on Attention"},{"key":"2530_CR36","unstructured":"Wang R, Dong Y, Arik S O, Yu R (2023) Koopman neural forecaster for time series with temporal distribution shifts. International Conference on Learning Representations"},{"key":"2530_CR37","unstructured":"Wang H, Peng J, Huang F, Wang J, Chen J, Xiao Y (2023) Micn: Multi-scale local and global context modeling for long-term series forecasting. In: The Eleventh International Conference on Learning Representations"},{"key":"2530_CR38","doi-asserted-by":"crossref","unstructured":"Han L, Ye H -J, Zhan D -C (2023) The capacity and robustness trade-off: Revisiting the channel independent strategy for multivariate time series forecasting. arXiv preprint arXiv:2304.05206","DOI":"10.1109\/TKDE.2024.3400008"},{"key":"2530_CR39","unstructured":"Jin M, Zhang Y, Chen W, Zhang K, Liang Y, Yang B, Wang J, Pan S, Wen Q (2024) Position: What can large language models tell us about time series analysis. In: Forty-first International Conference on Machine Learning"},{"key":"2530_CR40","unstructured":"Jin M, Wang S, Ma L, Chu Z, Zhang J Y, Shi X, Chen P -Y, Liang Y, Li Y -F, Pan S, Wen Q (2024) Time-LLM: Time series forecasting by reprogramming large language models. In: International Conference on Learning Representations (ICLR)"},{"key":"2530_CR41","doi-asserted-by":"crossref","unstructured":"Jia F, Wang K, Zheng Y, Cao D, Liu Y (2024) Gpt4mts: Prompt-based large language model for multimodal time-series forecasting. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, pp. 23343\u201323351","DOI":"10.1609\/aaai.v38i21.30383"},{"key":"2530_CR42","unstructured":"Zhang S, Wang S, Miao H, Chen H, Fan C, Zhang J (2024) Score-cdm: Score-weighted convolutional diffusion model for multivariate time series imputation. arXiv preprint arXiv:2405.13075"},{"key":"2530_CR43","doi-asserted-by":"crossref","unstructured":"Xu R, Miao H, Wang S, Yu P S, Wang J (2024) Pefad: a parameter-efficient federated framework for time series anomaly detection. In: Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 3621\u20133632","DOI":"10.1145\/3637528.3671753"},{"issue":"5","key":"2530_CR44","doi-asserted-by":"publisher","first-page":"1743","DOI":"10.1007\/s00778-024-00872-x","volume":"33","author":"X Wu","year":"2024","unstructured":"Wu X, Wu X, Yang B, Zhou L, Guo C, Qiu X, Hu J, Sheng Z, Jensen CS (2024) Autocts++: zero-shot joint neural architecture and hyperparameter search for correlated time series forecasting. VLDB J 33(5):1743\u20131770","journal-title":"VLDB J"},{"key":"2530_CR45","unstructured":"Liu Q, Liu X, Liu C, Wen Q, Liang Y (2024) Time-ffm: Towards lm-empowered federated foundation model for time series forecasting. arXiv preprint arXiv:2405.14252"},{"key":"2530_CR46","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez A N, Kaiser \u0141, Polosukhin I (2017) Attention is all you need. Advances in neural information processing systems 30"},{"key":"2530_CR47","doi-asserted-by":"crossref","unstructured":"Wen Q, Zhou T, Zhang C, Chen W, Ma Z, Yan J, Sun L (2022) Transformers in time series: A survey. arXiv preprint arXiv:2202.07125","DOI":"10.24963\/ijcai.2023\/759"},{"key":"2530_CR48","unstructured":"Subramanian S, Collobert R, Ranzato M, Boureau Y.-L (2020) Multi-scale transformer language models. arXiv preprint arXiv:2005.00581"},{"key":"2530_CR49","unstructured":"Shabani A, Abdi A, Meng L, Sylvain T (2023) Scaleformer: iterative multi-scale refining transformers for time series forecasting. International Conference on Learning Representations"},{"key":"2530_CR50","unstructured":"Yi K, Zhang Q, Fan W, Wang S, Wang P, He H, An N, Lian D, Cao L, Niu Z (2024) Frequency-domain mlps are more effective learners in time series forecasting. Adv Neural Inform Process Syst 36"},{"key":"2530_CR51","unstructured":"Liu Y, Li C, Wang J, Long M (2024) Koopa: Learning non-stationary time series dynamics with koopman predictors. Adv Neural Inform Process Syst 36"},{"key":"2530_CR52","unstructured":"Liu S, Yu H, Liao C, Li J, Lin W, Liu A X, Dustdar S (2021) Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting. In: International Conference on Learning Representations"},{"key":"2530_CR53","unstructured":"Kingma D P, Ba J (2014) Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980"},{"key":"2530_CR54","unstructured":"Liu Y, Wu H, Wang J, Long M (2022) Non-stationary transformers: Rethinking the stationarity in time series forecasting. NeurIPS"}],"container-title":["International Journal of Machine Learning and Cybernetics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-025-02530-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13042-025-02530-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-025-02530-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T09:41:38Z","timestamp":1765618898000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13042-025-02530-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,7]]},"references-count":54,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["2530"],"URL":"https:\/\/doi.org\/10.1007\/s13042-025-02530-2","relation":{},"ISSN":["1868-8071","1868-808X"],"issn-type":[{"type":"print","value":"1868-8071"},{"type":"electronic","value":"1868-808X"}],"subject":[],"published":{"date-parts":[[2025,2,7]]},"assertion":[{"value":"7 July 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 January 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 February 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no conflict of interest to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}