{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T10:08:54Z","timestamp":1775210934433,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":33,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,21]],"date-time":"2024-10-21T00:00:00Z","timestamp":1729468800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,21]]},"DOI":"10.1145\/3627673.3679931","type":"proceedings-article","created":{"date-parts":[[2024,10,20]],"date-time":"2024-10-20T19:34:11Z","timestamp":1729452851000},"page":"3757-3761","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":10,"title":["General Time Transformer: an Encoder-only Foundation Model for Zero-Shot Multivariate Time Series Forecasting"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0247-5355","authenticated-orcid":false,"given":"Cheng","family":"Feng","sequence":"first","affiliation":[{"name":"Siemens Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-9284-8371","authenticated-orcid":false,"given":"Long","family":"Huang","sequence":"additional","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-0249-0048","authenticated-orcid":false,"given":"Denis","family":"Krompass","sequence":"additional","affiliation":[{"name":"Siemens Technology, Munich, Germany"}]}],"member":"320","published-online":{"date-parts":[[2024,10,21]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Sebastian Pineda Arango, Shubham Kapoor, et al.","author":"Ansari Abdul Fatir","year":"2024","unstructured":"Abdul Fatir Ansari, Lorenzo Stella, Caner Turkmen, Xiyuan Zhang, Pedro Mercado, Huibin Shen, Oleksandr Shchur, Syama Sundar Rangapuram, Sebastian Pineda Arango, Shubham Kapoor, et al. 2024. Chronos: Learning the language of time series. arXiv preprint arXiv:2403.07815 (2024)."},{"key":"e_1_3_2_1_2_1","volume-title":"Forthcoming","author":"Borovykh Anastasia","year":"2018","unstructured":"Anastasia Borovykh, Sander Bohte, and Cornelis W Oosterlee. 2018. Dilated convolutional neural networks for time series forecasting. Journal of Computational Finance, Forthcoming (2018)."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.2307\/2985674"},{"key":"e_1_3_2_1_4_1","volume-title":"A decoder-only foundation model for time-series forecasting. arXiv preprint arXiv:2310.10688","author":"Das Abhimanyu","year":"2023","unstructured":"Abhimanyu Das, Weihao Kong, Rajat Sen, and Yichen Zhou. 2023. A decoder-only foundation model for time-series forecasting. arXiv preprint arXiv:2310.10688 (2023)."},{"key":"e_1_3_2_1_5_1","unstructured":"Alexey Dosovitskiy Lucas Beyer Alexander Kolesnikov Dirk Weissenborn Xiaohua Zhai Thomas Unterthiner Mostafa Dehghani Matthias Minderer Georg Heigold Sylvain Gelly et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)."},{"key":"e_1_3_2_1_6_1","volume-title":"Time series analysis by state space methods","author":"Durbin James","unstructured":"James Durbin and Siem Jan Koopman. 2012. Time series analysis by state space methods. Vol. 38. OUP Oxford."},{"key":"e_1_3_2_1_7_1","volume-title":"TSMixer: Lightweight MLP-Mixer Model for Multivariate Time Series Forecasting. arXiv preprint arXiv:2306.09364","author":"Ekambaram Vijay","year":"2023","unstructured":"Vijay Ekambaram, Arindam Jati, Nam Nguyen, Phanwadee Sinthong, and Jayant Kalagnanam. 2023. TSMixer: Lightweight MLP-Mixer Model for Multivariate Time Series Forecasting. arXiv preprint arXiv:2306.09364 (2023)."},{"key":"e_1_3_2_1_8_1","volume-title":"UniTS: Building a Unified Time Series Model. arXiv preprint arXiv:2403.00131","author":"Gao Shanghua","year":"2024","unstructured":"Shanghua Gao, Teddy Koker, Owen Queen, Thomas Hartvigsen, Theodoros Tsiligkaridis, and Marinka Zitnik. 2024. UniTS: Building a Unified Time Series Model. arXiv preprint arXiv:2403.00131 (2024)."},{"key":"e_1_3_2_1_9_1","volume-title":"arXiv preprint arXiv:2310.03589","author":"Garza Azul","year":"2023","unstructured":"Azul Garza and Max Mergenthaler-Canseco. 2023. TimeGPT-1. arXiv preprint arXiv:2310.03589 (2023)."},{"key":"e_1_3_2_1_10_1","volume-title":"Monash time series forecasting archive. arXiv preprint arXiv:2105.06643","author":"Godahewa Rakshitha","year":"2021","unstructured":"Rakshitha Godahewa, Christoph Bergmeir, Geoffrey I Webb, Rob J Hyndman, and Pablo Montero-Manso. 2021. Monash time series forecasting archive. arXiv preprint arXiv:2105.06643 (2021)."},{"key":"e_1_3_2_1_11_1","volume-title":"MOMENT: A Family of Open Time-series Foundation Models. In International Conference on Machine Learning.","author":"Goswami Mononito","year":"2024","unstructured":"Mononito Goswami, Konrad Szafer, Arjun Choudhry, Yifu Cai, Shuo Li, and Artur Dubrawski. 2024. MOMENT: A Family of Open Time-series Foundation Models. In International Conference on Machine Learning."},{"key":"e_1_3_2_1_12_1","volume-title":"Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415","author":"Hendrycks Dan","year":"2016","unstructured":"Dan Hendrycks and Kevin Gimpel. 2016. Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415 (2016)."},{"key":"e_1_3_2_1_13_1","volume-title":"Forecasting with exponential smoothing: the state space approach","author":"Hyndman Rob","unstructured":"Rob Hyndman, Anne B Koehler, J Keith Ord, and Ralph D Snyder. 2008. Forecasting with exponential smoothing: the state space approach. Springer Science & Business Media."},{"key":"e_1_3_2_1_14_1","volume-title":"International Conference on Learning Representations.","author":"Kim Taesung","year":"2021","unstructured":"Taesung Kim, Jinhee Kim, Yunwon Tae, Cheonbok Park, Jang-Ho Choi, and Jaegul Choo. 2021. Reversible instance normalization for accurate time-series forecasting against distribution shift. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_15_1","volume-title":"Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Advances in neural information processing systems","author":"Li Shiyang","year":"2019","unstructured":"Shiyang Li, Xiaoyong Jin, Yao Xuan, Xiyou Zhou, Wenhu Chen, Yu-Xiang Wang, and Xifeng Yan. 2019. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Advances in neural information processing systems, Vol. 32 (2019)."},{"key":"e_1_3_2_1_16_1","volume-title":"International conference on learning representations.","author":"Liu Shizhan","year":"2021","unstructured":"Shizhan Liu, Hang Yu, Cong Liao, Jianguo Li, Weiyao Lin, Alex X Liu, and Schahram Dustdar. 2021. Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting. In International conference on learning representations."},{"key":"e_1_3_2_1_17_1","volume-title":"itransformer: Inverted transformers are effective for time series forecasting. arXiv preprint arXiv:2310.06625","author":"Liu Yong","year":"2023","unstructured":"Yong Liu, Tengge Hu, Haoran Zhang, Haixu Wu, Shiyu Wang, Lintao Ma, and Mingsheng Long. 2023. itransformer: Inverted transformers are effective for time series forecasting. arXiv preprint arXiv:2310.06625 (2023)."},{"key":"e_1_3_2_1_18_1","volume-title":"Timer: Transformers for Time Series Analysis at Scale. arXiv preprint arXiv:2402.02368","author":"Liu Yong","year":"2024","unstructured":"Yong Liu, Haoran Zhang, Chenyu Li, Xiangdong Huang, Jianmin Wang, and Mingsheng Long. 2024. Timer: Transformers for Time Series Analysis at Scale. arXiv preprint arXiv:2402.02368 (2024)."},{"key":"e_1_3_2_1_19_1","volume-title":"Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101","author":"Loshchilov Ilya","year":"2017","unstructured":"Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)."},{"key":"e_1_3_2_1_20_1","volume-title":"A time series is worth 64 words: Long-term forecasting with transformers. arXiv preprint arXiv:2211.14730","author":"Nie Yuqi","year":"2022","unstructured":"Yuqi Nie, Nam H Nguyen, Phanwadee Sinthong, and Jayant Kalagnanam. 2022. A time series is worth 64 words: Long-term forecasting with transformers. arXiv preprint arXiv:2211.14730 (2022)."},{"key":"e_1_3_2_1_21_1","volume-title":"Arian Khorasani, George Adamopoulos, Rishika Bhagwatkar, Marin Bilovs, Hena Ghonia, Nadhir Vincent Hassen, Anderson Schneider, et al.","author":"Rasul Kashif","year":"2023","unstructured":"Kashif Rasul, Arjun Ashok, Andrew Robert Williams, Arian Khorasani, George Adamopoulos, Rishika Bhagwatkar, Marin Bilovs, Hena Ghonia, Nadhir Vincent Hassen, Anderson Schneider, et al. 2023. Lag-llama: Towards foundation models for time series forecasting. arXiv preprint arXiv:2310.08278 (2023)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2019.07.001"},{"key":"e_1_3_2_1_23_1","volume-title":"Attention is all you need. Advances in neural information processing systems","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems, Vol. 30 (2017)."},{"key":"e_1_3_2_1_24_1","volume-title":"Unified training of universal time series forecasting transformers. arXiv preprint arXiv:2402.02592","author":"Woo Gerald","year":"2024","unstructured":"Gerald Woo, Chenghao Liu, Akshat Kumar, Caiming Xiong, Silvio Savarese, and Doyen Sahoo. 2024. Unified training of universal time series forecasting transformers. arXiv preprint arXiv:2402.02592 (2024)."},{"key":"e_1_3_2_1_25_1","volume-title":"Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv preprint arXiv:2210.02186","author":"Wu Haixu","year":"2022","unstructured":"Haixu Wu, Tengge Hu, Yong Liu, Hang Zhou, Jianmin Wang, and Mingsheng Long. 2022. Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv preprint arXiv:2210.02186 (2022)."},{"key":"e_1_3_2_1_26_1","first-page":"22419","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","volume":"34","author":"Wu Haixu","year":"2021","unstructured":"Haixu Wu, Jiehui Xu, Jianmin Wang, and Mingsheng Long. 2021. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in Neural Information Processing Systems, Vol. 34 (2021), 22419--22430.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_27_1","volume-title":"AIM: Adapting Image Models for Efficient Video Action Recognition. In The Eleventh International Conference on Learning Representations.","author":"Yang Taojiannan","year":"2022","unstructured":"Taojiannan Yang, Yi Zhu, Yusheng Xie, Aston Zhang, Chen Chen, and Mu Li. 2022. AIM: Adapting Image Models for Efficient Video Action Recognition. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"e_1_3_2_1_29_1","volume-title":"The Eleventh International Conference on Learning Representations.","author":"Zhang Yunhao","year":"2022","unstructured":"Yunhao Zhang and Junchi Yan. 2022. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"e_1_3_2_1_31_1","volume-title":"International Conference on Machine Learning. PMLR, 27268--27286","author":"Zhou Tian","year":"2022","unstructured":"Tian Zhou, Ziqing Ma, Qingsong Wen, Xue Wang, Liang Sun, and Rong Jin. 2022. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In International Conference on Machine Learning. PMLR, 27268--27286."},{"key":"e_1_3_2_1_32_1","volume-title":"One Fits All: Power General Time Series Analysis by Pretrained LM. arXiv preprint arXiv:2302.11939","author":"Zhou Tian","year":"2023","unstructured":"Tian Zhou, Peisong Niu, Xue Wang, Liang Sun, and Rong Jin. 2023. One Fits All: Power General Time Series Analysis by Pretrained LM. arXiv preprint arXiv:2302.11939 (2023)."},{"key":"e_1_3_2_1_33_1","volume-title":"Vector autoregressive models for multivariate time series. Modeling financial time series with S-PLUS\u00ae","author":"Zivot Eric","year":"2006","unstructured":"Eric Zivot and Jiahui Wang. 2006. Vector autoregressive models for multivariate time series. Modeling financial time series with S-PLUS\u00ae (2006), 385--429."}],"event":{"name":"CIKM '24: The 33rd ACM International Conference on Information and Knowledge Management","location":"Boise ID USA","acronym":"CIKM '24","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval"]},"container-title":["Proceedings of the 33rd ACM International Conference on Information and Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3627673.3679931","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3627673.3679931","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:58:09Z","timestamp":1750294689000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3627673.3679931"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,21]]},"references-count":33,"alternative-id":["10.1145\/3627673.3679931","10.1145\/3627673"],"URL":"https:\/\/doi.org\/10.1145\/3627673.3679931","relation":{},"subject":[],"published":{"date-parts":[[2024,10,21]]},"assertion":[{"value":"2024-10-21","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}