{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T18:40:10Z","timestamp":1772563210889,"version":"3.50.1"},"reference-count":34,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T00:00:00Z","timestamp":1751241600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T00:00:00Z","timestamp":1751241600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100006190","name":"Research and Development","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100006190","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,6,30]]},"DOI":"10.1109\/ijcnn64981.2025.11229402","type":"proceedings-article","created":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T18:46:15Z","timestamp":1763145975000},"page":"1-8","source":"Crossref","is-referenced-by-count":3,"title":["MTS-Mixers: Multivariate Time Series Forecasting via Factorized Temporal and Channel Mixing"],"prefix":"10.1109","author":[{"given":"Zhe","family":"Li","sequence":"first","affiliation":[{"name":"Harbin Institute of Technology, Shenzhen,Shenzhen,China"}]},{"given":"Xuanxuan","family":"Li","sequence":"additional","affiliation":[{"name":"Harbin Institute of Technology, Shenzhen,Shenzhen,China"}]},{"given":"Zhongwen","family":"Rao","sequence":"additional","affiliation":[{"name":"Huawei Technologies Ltd,Shenzhen,China"}]},{"given":"Lujia","family":"Pan","sequence":"additional","affiliation":[{"name":"Huawei Technologies Ltd,Shenzhen,China"}]},{"given":"Zenglin","family":"Xu","sequence":"additional","affiliation":[{"name":"Fudan University,Shanghai,China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.3390\/s20051399"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1038\/s41597-020-0548-x"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.3141\/1748-12"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3210006"},{"key":"ref5","article-title":"An empirical evaluation of generic convolutional and recurrent networks for sequence modeling","author":"Bai","year":"2018"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1016\/S0925-2312(01)00702-0"},{"key":"ref7","article-title":"Attention is all you need","author":"Vaswani","year":"2017"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"ref9","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","author":"Wu","year":"2021","journal-title":"Neural Information Processing Systems"},{"key":"ref10","article-title":"Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting","volume-title":"International Conference on Learning Representations","author":"Liu"},{"key":"ref11","article-title":"Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting","author":"Zhou","year":"2022"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11635"},{"key":"ref14","article-title":"Cdsa: Cross-dimensional self-attention for multivariate, geo-tagged time series imputation","author":"Ma","year":"2019"},{"key":"ref15","article-title":"Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting","author":"Li","year":"2019"},{"key":"ref16","article-title":"Generating long sequences with sparse transformers","author":"Child","year":"2019"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2023\/241"},{"key":"ref18","article-title":"A time series is worth 64 words: Long-term forecasting with transformers","volume-title":"The Eleventh International Conference on Learning Representations","author":"Nie"},{"key":"ref19","article-title":"itrans-former: Inverted transformers are effective for time series forecasting","volume-title":"The Twelfth International Conference on Learning Representations","author":"Liu"},{"key":"ref20","article-title":"Enhancing multivariate time series forecasting with mutual information-driven cross-variable and temporal modeling","author":"Qi","year":"2024"},{"key":"ref21","article-title":"Etsformer: Exponential smoothing transformers for time-series forecasting","author":"Woo","year":"2022"},{"key":"ref22","article-title":"Pdetime: Rethinking long-term multivariate time series forecasting from the perspective of partial differential equations","author":"Qi","year":"2024"},{"key":"ref23","article-title":"Revisiting long-term time series forecasting: An investigation on linear mapping","author":"Li","year":"2023"},{"key":"ref24","article-title":"Synthesizer: Rethinking self-attention in transformer models","volume-title":"International Conference on Machine Learning","author":"Tay"},{"key":"ref25","article-title":"Mlp-mixer: An all-mlp architecture for vision","author":"Tolstikhin","year":"2021","journal-title":"Neural Information Processing Systems"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.naacl-main.319"},{"key":"ref27","article-title":"Metaformer: A unified meta framework for fine-grained recognition","author":"Diao","year":"2022"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.6028\/NIST.IR.6131"},{"key":"ref29","article-title":"Is attention better than matrix decomposition?","author":"Geng","year":"2021"},{"key":"ref30","article-title":"Timesnet: Temporal 2d-variation modeling for general time series analysis","volume-title":"The Eleventh International Conference on Learning Representations","author":"Wu"},{"key":"ref31","article-title":"SCINet: Time series modeling and forecasting with sample convolution and interaction","volume-title":"Advances in Neural Information Processing Systems","author":"Liu","year":"2022"},{"key":"ref32","article-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2015","journal-title":"CoRR"},{"key":"ref33","article-title":"Pytorch: An imperative style, high-performance deep learning library","volume-title":"NeurIPS","author":"Paszke"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00591"}],"event":{"name":"2025 International Joint Conference on Neural Networks (IJCNN)","location":"Rome, Italy","start":{"date-parts":[[2025,6,30]]},"end":{"date-parts":[[2025,7,5]]}},"container-title":["2025 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11227166\/11227148\/11229402.pdf?arnumber=11229402","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T07:31:13Z","timestamp":1763191873000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11229402\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,30]]},"references-count":34,"URL":"https:\/\/doi.org\/10.1109\/ijcnn64981.2025.11229402","relation":{},"subject":[],"published":{"date-parts":[[2025,6,30]]}}}