{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T22:50:36Z","timestamp":1776207036395,"version":"3.50.1"},"reference-count":44,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100012253","name":"Guangxi University","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100012253","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004607","name":"Guangxi Natural Science Foundation","doi-asserted-by":"publisher","award":["2025GXNSFAA069359"],"award-info":[{"award-number":["2025GXNSFAA069359"]}],"id":[{"id":"10.13039\/501100004607","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Engineering Applications of Artificial Intelligence"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1016\/j.engappai.2025.113610","type":"journal-article","created":{"date-parts":[[2025,12,26]],"date-time":"2025-12-26T10:13:16Z","timestamp":1766743996000},"page":"113610","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":1,"special_numbering":"PB","title":["A dual-branch multi-scale encoding and fusion model for multivariate time series forecasting"],"prefix":"10.1016","volume":"166","author":[{"given":"Jiachao","family":"Li","sequence":"first","affiliation":[]},{"given":"Mengxiao","family":"Yin","sequence":"additional","affiliation":[]},{"given":"Junyuan","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Tao","family":"Luo","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.engappai.2025.113610_b1","series-title":"Time Series Analysis: Forecasting and Control","author":"Box","year":"1976"},{"key":"10.1016\/j.engappai.2025.113610_b2","series-title":"Tsmixer: An all-mlp architecture for time series forecasting","author":"Chen","year":"2023"},{"key":"10.1016\/j.engappai.2025.113610_b3","series-title":"Pathformer: Multi-scale transformers with adaptive pathways for time series forecasting","author":"Chen","year":"2024"},{"key":"10.1016\/j.engappai.2025.113610_b4","series-title":"Empirical evaluation of gated recurrent neural networks on sequence modeling","author":"Chung","year":"2014"},{"key":"10.1016\/j.engappai.2025.113610_b5","series-title":"An image is worth 16x16 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020"},{"key":"10.1016\/j.engappai.2025.113610_b6","series-title":"Forecasting: Principles and Practice","author":"Hyndman","year":"2018"},{"key":"10.1016\/j.engappai.2025.113610_b7","doi-asserted-by":"crossref","first-page":"1","DOI":"10.18637\/jss.v027.i03","article-title":"Automatic time series forecasting: The forecast package for R","volume":"27","author":"Hyndman","year":"2008","journal-title":"J. Stat. Softw."},{"key":"10.1016\/j.engappai.2025.113610_b8","series-title":"2022 IEEE 38th International Conference on Data Engineering","first-page":"1342","article-title":"Anomaly detection in time series with robust variational quasi-recurrent autoencoders","author":"Kieu","year":"2022"},{"key":"10.1016\/j.engappai.2025.113610_b9","unstructured":"Kim, T., Kim, J., Tae, Y., Park, C., Choi, J.-H., Choo, J., 2021. Reversible instance normalization for accurate time-series forecasting against distribution shift. In: International Conference on Learning Representations."},{"key":"10.1016\/j.engappai.2025.113610_b10","doi-asserted-by":"crossref","first-page":"114180","DOI":"10.52202\/079017-3627","article-title":"Are self-attentions effective for time series forecasting?","volume":"37","author":"Kim","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"6677","key":"10.1016\/j.engappai.2025.113610_b11","doi-asserted-by":"crossref","first-page":"1416","DOI":"10.1126\/science.adi2336","article-title":"Learning skillful medium-range global weather forecasting","volume":"382","author":"Lam","year":"2023","journal-title":"Science"},{"key":"10.1016\/j.engappai.2025.113610_b12","doi-asserted-by":"crossref","unstructured":"Li, H., Shao, J., Liao, K., Tang, M., 2022. Do simpler statistical methods perform better in multivariate long sequence time-series forecasting?. In: Proceedings of the 31st ACM International Conference on Information & Knowledge Management. pp. 4168\u20134172.","DOI":"10.1145\/3511808.3557585"},{"key":"10.1016\/j.engappai.2025.113610_b13","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2023.109423","article-title":"Dynamic graph structure learning for multivariate time series forecasting","volume":"138","author":"Li","year":"2023","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.engappai.2025.113610_b14","series-title":"Itransformer: Inverted transformers are effective for time series forecasting","author":"Liu","year":"2023"},{"key":"10.1016\/j.engappai.2025.113610_b15","series-title":"Timer-xl: Long-context transformers for unified time series forecasting","author":"Liu","year":"2024"},{"key":"10.1016\/j.engappai.2025.113610_b16","doi-asserted-by":"crossref","first-page":"9881","DOI":"10.52202\/068431-0718","article-title":"Non-stationary transformers: Exploring the stationarity in time series forecasting","volume":"35","author":"Liu","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.engappai.2025.113610_b17","series-title":"# Placeholder_Parent_Metadata_Value#","article-title":"Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting","author":"Liu","year":"2022"},{"key":"10.1016\/j.engappai.2025.113610_b18","series-title":"Fixing weight decay regularization in adam","author":"Loshchilov","year":"2017"},{"key":"10.1016\/j.engappai.2025.113610_b19","series-title":"2024 IEEE 40th International Conference on Data Engineering","first-page":"1050","article-title":"A unified replay-based continuous learning framework for spatio-temporal prediction on streaming data","author":"Miao","year":"2024"},{"issue":"1","key":"10.1016\/j.engappai.2025.113610_b20","doi-asserted-by":"crossref","first-page":"1565","DOI":"10.1038\/s41598-024-82417-4","article-title":"A multiscale model for multivariate time series forecasting","volume":"15","author":"Naghashi","year":"2025","journal-title":"Sci. Rep."},{"key":"10.1016\/j.engappai.2025.113610_b21","series-title":"A time series is worth 64 words: Long-term forecasting with transformers","author":"Nie","year":"2022"},{"key":"10.1016\/j.engappai.2025.113610_b22","series-title":"TEA: Trajectory encoding augmentation for robust and transferable policies in offline reinforcement learning","author":"Ormanc\u0131","year":"2024"},{"key":"10.1016\/j.engappai.2025.113610_b23","article-title":"Pytorch: An imperative style, high-performance deep learning library","volume":"32","author":"Paszke","year":"2019","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"6","key":"10.1016\/j.engappai.2025.113610_b24","doi-asserted-by":"crossref","first-page":"H2039","DOI":"10.1152\/ajpheart.2000.278.6.H2039","article-title":"Physiological time-series analysis using approximate entropy and sample entropy","volume":"278","author":"Richman","year":"2000","journal-title":"Am. J. Physiol.-Heart Circ. Physiol."},{"key":"10.1016\/j.engappai.2025.113610_b25","series-title":"Scaleformer: Iterative multi-scale refining transformers for time series forecasting","author":"Shabani","year":"2022"},{"key":"10.1016\/j.engappai.2025.113610_b26","first-page":"24261","article-title":"Mlp-mixer: An all-mlp architecture for vision","volume":"34","author":"Tolstikhin","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.engappai.2025.113610_b27","article-title":"Attention is all you need","volume":"30","author":"Vaswani","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"5","key":"10.1016\/j.engappai.2025.113610_b28","doi-asserted-by":"crossref","first-page":"3123","DOI":"10.1109\/TPAMI.2023.3341806","article-title":"Crossformer++: A versatile vision transformer hinging on cross-scale attention","volume":"46","author":"Wang","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.engappai.2025.113610_b29","doi-asserted-by":"crossref","unstructured":"Wang, H., Mo, Y., Xiang, K., Yin, N., Dai, H., Li, B., Fan, S., 2025. CSformer: Combining Channel Independence and Mixing for Robust Multivariate Time Series Forecasting. In: Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 39, pp. 21090\u201321098.","DOI":"10.1609\/aaai.v39i20.35406"},{"key":"10.1016\/j.engappai.2025.113610_b30","first-page":"469","article-title":"Timexer: Empowering transformers for time series forecasting with exogenous variables","volume":"37","author":"Wang","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.engappai.2025.113610_b31","series-title":"Timemixer: Decomposable multiscale mixing for time series forecasting","author":"Wang","year":"2024"},{"key":"10.1016\/j.engappai.2025.113610_b32","doi-asserted-by":"crossref","unstructured":"Wang, W., Xie, E., Li, X., Fan, D.-P., Song, K., Liang, D., Lu, T., Luo, P., Shao, L., 2021. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 568\u2013578.","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"10.1016\/j.engappai.2025.113610_b33","series-title":"Timesnet: Temporal 2d-variation modeling for general time series analysis","author":"Wu","year":"2022"},{"key":"10.1016\/j.engappai.2025.113610_b34","doi-asserted-by":"crossref","unstructured":"Wu, Z., Pan, S., Long, G., Jiang, J., Chang, X., Zhang, C., 2020. Connecting the dots: Multivariate time series forecasting with graph neural networks. In: Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. pp. 753\u2013763.","DOI":"10.1145\/3394486.3403118"},{"key":"10.1016\/j.engappai.2025.113610_b35","first-page":"22419","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","volume":"34","author":"Wu","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.engappai.2025.113610_b36","doi-asserted-by":"crossref","DOI":"10.1109\/TBDATA.2025.3594294","article-title":"Knowledge aggregation transformer network for multivariate time series classification","author":"Xiao","year":"2025","journal-title":"IEEE Trans. Big Data"},{"key":"10.1016\/j.engappai.2025.113610_b37","series-title":"Spatial-temporal transformer networks for traffic flow forecasting","author":"Xu","year":"2020"},{"issue":"1","key":"10.1016\/j.engappai.2025.113610_b38","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3691338","article-title":"Deep learning for time series anomaly detection: A survey","volume":"57","author":"Zamanzadeh Darban","year":"2024","journal-title":"ACM Comput. Surv."},{"issue":"28","key":"10.1016\/j.engappai.2025.113610_b39","doi-asserted-by":"crossref","first-page":"23533","DOI":"10.1007\/s00521-025-11538-2","article-title":"An interpretable information fusion approach to impute meteorological missing values toward cross-domain intelligent forecasting","volume":"37","author":"Zarchi","year":"2025","journal-title":"Neural Comput. Appl."},{"key":"10.1016\/j.engappai.2025.113610_b40","doi-asserted-by":"crossref","unstructured":"Zeng, A., Chen, M., Zhang, L., Xu, Q., 2023a. Are transformers effective for time series forecasting?. In: Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 37, pp. 11121\u201311128.","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"10.1016\/j.engappai.2025.113610_b41","series-title":"Financial time series forecasting using cnn and transformer","author":"Zeng","year":"2023"},{"key":"10.1016\/j.engappai.2025.113610_b42","unstructured":"Zhang, Y., Yan, J., 2023. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In: The Eleventh International Conference on Learning Representations."},{"key":"10.1016\/j.engappai.2025.113610_b43","series-title":"International Conference on Machine Learning","first-page":"27268","article-title":"Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting","author":"Zhou","year":"2022"},{"key":"10.1016\/j.engappai.2025.113610_b44","doi-asserted-by":"crossref","unstructured":"Zhou, H., Zhang, S., Peng, J., Zhang, S., Li, J., Xiong, H., Zhang, W., 2021. Informer: Beyond efficient transformer for long sequence time-series forecasting. In: Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 35, pp. 11106\u201311115.","DOI":"10.1609\/aaai.v35i12.17325"}],"container-title":["Engineering Applications of Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0952197625036425?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0952197625036425?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T14:50:10Z","timestamp":1773931810000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0952197625036425"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2]]},"references-count":44,"alternative-id":["S0952197625036425"],"URL":"https:\/\/doi.org\/10.1016\/j.engappai.2025.113610","relation":{},"ISSN":["0952-1976"],"issn-type":[{"value":"0952-1976","type":"print"}],"subject":[],"published":{"date-parts":[[2026,2]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"A dual-branch multi-scale encoding and fusion model for multivariate time series forecasting","name":"articletitle","label":"Article Title"},{"value":"Engineering Applications of Artificial Intelligence","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.engappai.2025.113610","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"113610"}}