{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T20:10:14Z","timestamp":1771877414966,"version":"3.50.1"},"reference-count":61,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100007246","name":"YOK","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100007246","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Neural Networks"],"published-print":{"date-parts":[[2026,5]]},"DOI":"10.1016\/j.neunet.2025.108495","type":"journal-article","created":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T16:12:16Z","timestamp":1766074336000},"page":"108495","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":1,"special_numbering":"C","title":["CANet: ChronoAdaptive network for enhanced long-term time series forecasting under non-stationarity"],"prefix":"10.1016","volume":"197","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-1016-2449","authenticated-orcid":false,"given":"Mert","family":"Sonmezer","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6132-6739","authenticated-orcid":false,"given":"Seyda","family":"Ertekin","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.neunet.2025.108495_bib0001","series-title":"Advances in neural information processing systems","article-title":"Invariance principle meets information bottleneck for out-of-distribution generalization","volume":"vol. 5","author":"Ahuja","year":"2021"},{"key":"10.1016\/j.neunet.2025.108495_bib0002","series-title":"Proceedings of the ACM SIGKDD international conference on knowledge discovery and data mining","article-title":"Optuna: A next-generation hyperparameter optimization framework","author":"Akiba","year":"2019"},{"key":"10.1016\/j.neunet.2025.108495_bib0003","doi-asserted-by":"crossref","DOI":"10.2307\/2988091","article-title":"Time-series. 2nd edn","volume":"25","author":"Anderson","year":"1976","journal-title":"The Statistician"},{"key":"10.1016\/j.neunet.2025.108495_bib0004","unstructured":"Ba, J. L., Kiros, J. R., & Hinton, G. E. (2016). Layer normalization. cite arxiv: 1607.06450 http:\/\/arxiv.org\/abs\/1607.06450."},{"key":"10.1016\/j.neunet.2025.108495_bib0005","unstructured":"Bai, S., Kolter, J. Z., & Koltun, V. (2018). An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv: 1803.01271."},{"key":"10.1016\/j.neunet.2025.108495_bib0006","doi-asserted-by":"crossref","DOI":"10.2307\/2985674","article-title":"Some recent advances in forecasting and control","volume":"17","author":"Box","year":"1968","journal-title":"Applied Statistics"},{"key":"10.1016\/j.neunet.2025.108495_bib0007","unstructured":"Cheng, M., Liu, Q., Liu, Z., Zhang, H., Zhang, R., & Chen, E. (2023). TimeMAE: Self-supervised representations of time series with decoupled masked autoencoders. arXiv preprint arXiv: 2303.00320."},{"key":"10.1016\/j.neunet.2025.108495_bib0008","series-title":"Proceedings of the ACM SIGKDD international conference on knowledge discovery and data mining","article-title":"ST-norm: Spatial and temporal normalization for multi-variate time series forecasting","author":"Deng","year":"2021"},{"key":"10.1016\/j.neunet.2025.108495_bib0009","series-title":"The twelfth international conference on learning representations","article-title":"ModernTCN: A modern pure convolution structure for general time series analysis","author":"Donghao","year":"2024"},{"key":"10.1016\/j.neunet.2025.108495_bib0010","series-title":"International conference on machine learning","article-title":"TSLANet: Rethinking transformers for time series representation learning","author":"Eldele","year":"2024"},{"key":"10.1016\/j.neunet.2025.108495_bib0011","doi-asserted-by":"crossref","DOI":"10.2307\/2171846","article-title":"Efficient tests for an autoregressive unit root","volume":"64","author":"Elliott","year":"1996","journal-title":"Econometrica : Journal of the Econometric Society"},{"key":"10.1016\/j.neunet.2025.108495_bib0012","doi-asserted-by":"crossref","unstructured":"Fan, W., Wang, P., Wang, D., Wang, D., Zhou, Y., & Fu, Y. (2023). Dish-TS: A general paradigm for alleviating distribution shift in time series forecasting. https:\/\/arxiv.org\/abs\/2302.14829.","DOI":"10.1609\/aaai.v37i6.25914"},{"key":"10.1016\/j.neunet.2025.108495_bib0013","series-title":"Advances in neural information processing systems","article-title":"Unsupervised scalable representation learning for multivariate time series","volume":"vol. 32","author":"Franceschi","year":"2019"},{"key":"10.1016\/j.neunet.2025.108495_bib0014","series-title":"Proceedings of the 42nd international conference on machine learning","first-page":"19172","article-title":"Time series representations with hard-coded invariances","volume":"vol. 267","author":"Germain","year":"2025"},{"key":"10.1016\/j.neunet.2025.108495_bib0015","series-title":"2024\u202fIEEE 40th international conference on data engineering workshops (ICDEW)","first-page":"167","article-title":"Linear-trend normalization for multivariate subsequence similarity search","author":"Germain","year":"2024"},{"key":"10.1016\/j.neunet.2025.108495_bib0016","article-title":"Time series analysis: Forecasting and control","volume":"14","author":"Geurts","year":"1977","journal-title":"Journal of Marketing Research"},{"key":"10.1016\/j.neunet.2025.108495_bib0017","series-title":"Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics)","article-title":"ShapeWordNet: An interpretable shapelet neural network for physiological signal classification","volume":"vol. 13946 LNCS","author":"He","year":"2023"},{"key":"10.1016\/j.neunet.2025.108495_bib0018","doi-asserted-by":"crossref","DOI":"10.1162\/neco.1997.9.8.1735","article-title":"Long short-term memory","volume":"9","author":"Hochreiter","year":"1997","journal-title":"Neural Computation"},{"key":"10.1016\/j.neunet.2025.108495_bib0019","article-title":"Arbitrary style transfer in real-time with adaptive instance normalization","volume":"vol. 2017-October","author":"Huang","year":"2017","journal-title":"Proceedings of the IEEE International Conference on Computer Vision"},{"key":"10.1016\/j.neunet.2025.108495_bib0020","first-page":"66","article-title":"Forecasting: Principles and practice, 3rd edition","volume":"22","author":"Hyndman","year":"2021","journal-title":"International Journal of Forecasting"},{"key":"10.1016\/j.neunet.2025.108495_bib0021","series-title":"32nd international conference on machine learning, ICML 2015","article-title":"Batch normalization: Accelerating deep network training by reducing internal covariate shift","volume":"vol. 1","author":"Ioffe","year":"2015"},{"key":"10.1016\/j.neunet.2025.108495_bib0022","series-title":"Iclr 2022 - 10th international conference on learning representations","article-title":"Reversible instance normalization for accurate time-series forecasting against distribution shift","author":"Kim","year":"2022"},{"key":"10.1016\/j.neunet.2025.108495_bib0023","series-title":"8th international conference on learning representations, ICLR 2020","article-title":"Reformer: The efficient transformer","author":"Kitaev","year":"2020"},{"key":"10.1016\/j.neunet.2025.108495_bib0024","doi-asserted-by":"crossref","DOI":"10.1007\/s10472-019-09683-1","article-title":"Discrepancy-based theory and algorithms for forecasting non-stationary time series","volume":"88","author":"Kuznetsov","year":"2020","journal-title":"Annals of Mathematics and Artificial Intelligence"},{"key":"10.1016\/j.neunet.2025.108495_bib0025","series-title":"41st international ACM SIGIR conference on research and development in information retrieval, SIGIR 2018","article-title":"Modeling long- and short-term temporal patterns with deep neural networks","author":"Lai","year":"2018"},{"key":"10.1016\/j.neunet.2025.108495_bib0026","series-title":"Proceedings of the IEEE international conference on computer vision","article-title":"Deeper, broader and artier domain generalization","volume":"vol. 2017-October","author":"Li","year":"2017"},{"key":"10.1016\/j.neunet.2025.108495_bib0027","series-title":"ICLR 2022 - 10th international conference on learning representations","article-title":"Uniformer: Unified transformer for efficient spatiotemporal representation learning","author":"Li","year":"2022"},{"key":"10.1016\/j.neunet.2025.108495_bib0028","unstructured":"Liu, H., Kamarthi, H., Kong, L., Zhao, Z., Zhang, C., & Prakash, B. A. (2024a). Time-series forecasting for out-of-distribution generalization using invariant learning. https:\/\/arxiv.org\/abs\/2406.09130."},{"key":"10.1016\/j.neunet.2025.108495_bib0029","series-title":"Advances in neural information processing systems","article-title":"SCINet: Time series modeling and forecasting with sample convolution and interaction","volume":"vol. 35","author":"Liu","year":"2022"},{"key":"10.1016\/j.neunet.2025.108495_bib0030","series-title":"ICLR 2022 - 10th international conference on learning representations","article-title":"T-WaveNet: A tree-structured wavelet neural network for time series signal analysis","author":"Liu","year":"2022"},{"key":"10.1016\/j.neunet.2025.108495_bib0031","series-title":"ICASSP 2024 - 2024 IEEE international conference on acoustics, speech and signal processing (icassp)","first-page":"5960","article-title":"WFTNet: Exploiting global and local periodicity in long-term time series forecasting","author":"Liu","year":"2024"},{"key":"10.1016\/j.neunet.2025.108495_bib0032","series-title":"The twelfth international conference on learning representations","article-title":"Itransformer: Inverted transformers are effective for time series forecasting","author":"Liu","year":"2024"},{"key":"10.1016\/j.neunet.2025.108495_bib0033","unstructured":"Liu, Y., Li, C., Wang, J., & Long, M. (2023a). Koopa: Learning non-stationary time series dynamics with koopman predictors. https:\/\/arxiv.org\/abs\/2305.18803."},{"key":"10.1016\/j.neunet.2025.108495_bib0034","series-title":"Advances in neural information processing systems","article-title":"Non-stationary transformers: Exploring the stationarity in time series forecasting","volume":"vol. 35","author":"Liu","year":"2022"},{"key":"10.1016\/j.neunet.2025.108495_bib0035","series-title":"Advances in neural information processing systems","article-title":"Adaptive normalization for non-stationary time series forecasting: A temporal slice perspective","volume":"vol. 36","author":"Liu","year":"2023"},{"key":"10.1016\/j.neunet.2025.108495_bib0036","unstructured":"Maddix, D. C., Wang, Y., & Smola, A. (2018). Deep factors with gaussian processes for forecasting, 10. arXiv preprint arXiv: 1812.00098."},{"key":"10.1016\/j.neunet.2025.108495_bib0037","series-title":"The eleventh international conference on learning representations","article-title":"A time series is worth 64 words: Long-term forecasting with transformers","author":"Nie","year":"2023"},{"key":"10.1016\/j.neunet.2025.108495_bib0038","series-title":"Proceedings of the international joint conference on neural networks","article-title":"Adaptive normalization: A novel data normalization approach for non-stationary time series","author":"Ogasawara","year":"2010"},{"key":"10.1016\/j.neunet.2025.108495_bib0039","series-title":"8th international conference on learning representations, ICLR 2020","article-title":"N-beats: neural basis expansion analysis for interpretable time series forecasting","author":"Oreshkin","year":"2020"},{"key":"10.1016\/j.neunet.2025.108495_bib0040","doi-asserted-by":"crossref","unstructured":"Pan, S. J., & Yang, Q. (2010). A survey on transfer learning. 10.1109\/TKDE.2009.191.","DOI":"10.1109\/TKDE.2009.191"},{"key":"10.1016\/j.neunet.2025.108495_bib0041","series-title":"Advances in neural information processing systems","article-title":"ShapeShifter: A parameter-efficient transformer using factorized reshaped matrices","volume":"vol. 2","author":"Panahi","year":"2021"},{"key":"10.1016\/j.neunet.2025.108495_bib0042","doi-asserted-by":"crossref","DOI":"10.1109\/TNNLS.2019.2944933","article-title":"Deep adaptive input normalization for time series forecasting","volume":"31","author":"Passalis","year":"2020","journal-title":"IEEE Transactions on Neural Networks and Learning Systems"},{"key":"10.1016\/j.neunet.2025.108495_bib0043","series-title":"Advances in neural information processing systems","article-title":"PyTorch: An imperative style, high-performance deep learning library","volume":"vol. 32","author":"Paszke","year":"2019"},{"key":"10.1016\/j.neunet.2025.108495_bib0044","series-title":"Advances in neural information processing systems","article-title":"Think globally, act locally: A deep neural network approach to high-dimensional time series forecasting","volume":"vol. 32","author":"Rajat","year":"2019"},{"key":"10.1016\/j.neunet.2025.108495_bib0045","series-title":"Advances in neural information processing systems","article-title":"Deep state space models for time series forecasting","volume":"vol. 2018-December","author":"Rangapuram","year":"2018"},{"key":"10.1016\/j.neunet.2025.108495_bib0046","doi-asserted-by":"crossref","DOI":"10.1016\/j.ijforecast.2019.07.001","article-title":"DeepAR: Probabilistic forecasting with autoregressive recurrent networks","volume":"36","author":"Salinas","year":"2020","journal-title":"International Journal of Forecasting"},{"key":"10.1016\/j.neunet.2025.108495_bib0047","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2023.110666","article-title":"FDNet: Focal decomposed network for efficient, robust and practical time series forecasting","volume":"275","author":"Shen","year":"2023","journal-title":"Knowledge-Based Systems"},{"key":"10.1016\/j.neunet.2025.108495_bib0048","unstructured":"Ulyanov, D., Vedaldi, A., & Lempitsky, V. (2017). Instance normalization: The missing ingredient for fast stylization. arXiv preprint arXiv: 1607.08022."},{"key":"10.1016\/j.neunet.2025.108495_bib0049","series-title":"Advances in neural information processing systems","article-title":"Attention is all you need","volume":"vol. 2017-December","author":"Vaswani","year":"2017"},{"key":"10.1016\/j.neunet.2025.108495_bib0050","series-title":"11th international conference on learning representations, ICLR 2023","article-title":"MICN: Multi-scale local and global context modeling for long-term series forecasting","author":"Wang","year":"2023"},{"key":"10.1016\/j.neunet.2025.108495_bib0051","series-title":"IJCAI international joint conference on artificial intelligence","article-title":"Transformers in time series: A survey","volume":"vol. 2023-August","author":"Wen","year":"2023"},{"key":"10.1016\/j.neunet.2025.108495_bib0052","unstructured":"Wen, R., Torkkola, K., Narayanaswamy, B., & Madeka, D. (2017). A multi-horizon quantile recurrent forecaster. arXiv preprint arXiv: 1711.11053."},{"key":"10.1016\/j.neunet.2025.108495_bib0053","unstructured":"Wu, H., Hu, T., Liu, Y., Zhou, H., Wang, J., & Long, M. (2023). TimesNet: Temporal 2D-variation modeling for general time series analysis. https:\/\/arxiv.org\/abs\/2210.02186."},{"key":"10.1016\/j.neunet.2025.108495_bib0054","series-title":"Advances in neural information processing systems","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","volume":"vol. 27","author":"Wu","year":"2021"},{"key":"10.1016\/j.neunet.2025.108495_bib0055","doi-asserted-by":"crossref","unstructured":"Ye, W., Deng, S., Zou, Q., & Gui, N. (2024). Frequency adaptive normalization for non-stationary time series forecasting. https:\/\/arxiv.org\/abs\/2409.20371.","DOI":"10.52202\/079017-0985"},{"key":"10.1016\/j.neunet.2025.108495_bib0056","unstructured":"Yu, R., Zheng, S., Anandkumar, A., & Yue, Y. (2017). Long-term forecasting using tensor-train RNNs. arXiv."},{"key":"10.1016\/j.neunet.2025.108495_bib0057","series-title":"Proceedings of the 37th AAAI conference on artificial intelligence, AAAI 2023","article-title":"Are transformers effective for time series forecasting?","volume":"vol. 37","author":"Zeng","year":"2023"},{"key":"10.1016\/j.neunet.2025.108495_bib0058","unstructured":"Zhang, T., Zhang, Y., Cao, W., Bian, J., Yi, X., Zheng, S., & Li, J. (2022). Less is more: Fast multivariate time series forecasting with light sampling-oriented MLP structures. https:\/\/arxiv.org\/abs\/2207.01186."},{"key":"10.1016\/j.neunet.2025.108495_bib0059","series-title":"Proceedings of the 27th international conference on artificial intelligence and statistics","first-page":"4222","article-title":"Multi-resolution time-series transformer for long-term forecasting","volume":"vol. 238","author":"Zhang","year":"2024"},{"key":"10.1016\/j.neunet.2025.108495_bib0060","series-title":"35h AAAI conference on artificial intelligence, AAAI 2021","article-title":"Informer: Beyond efficient transformer for long sequence time-series forecasting","volume":"vol. 12B","author":"Zhou","year":"2021"},{"key":"10.1016\/j.neunet.2025.108495_bib0061","series-title":"Proceedings of machine learning research","article-title":"FEDFormer: Frequency enhanced decomposed transformer for long-term series forecasting","volume":"vol. 162","author":"Zhou","year":"2022"}],"container-title":["Neural Networks"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0893608025013760?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0893608025013760?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T19:54:50Z","timestamp":1771876490000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0893608025013760"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,5]]},"references-count":61,"alternative-id":["S0893608025013760"],"URL":"https:\/\/doi.org\/10.1016\/j.neunet.2025.108495","relation":{},"ISSN":["0893-6080"],"issn-type":[{"value":"0893-6080","type":"print"}],"subject":[],"published":{"date-parts":[[2026,5]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"CANet: ChronoAdaptive network for enhanced long-term time series forecasting under non-stationarity","name":"articletitle","label":"Article Title"},{"value":"Neural Networks","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.neunet.2025.108495","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"108495"}}