{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T08:12:08Z","timestamp":1769847128318,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":44,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,8,3]]},"DOI":"10.1145\/3711896.3737157","type":"proceedings-article","created":{"date-parts":[[2025,8,3]],"date-time":"2025-08-03T21:03:27Z","timestamp":1754255007000},"page":"1987-1998","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["TimeCapsule: Solving the Jigsaw Puzzle of Long-Term Time Series Forecasting with Compressed Predictive Representations"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2566-9427","authenticated-orcid":false,"given":"Yihang","family":"Lu","sequence":"first","affiliation":[{"name":"Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei, Anhui, China and University of Science and Technology of China, Hefei, Anhui, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-3679-4390","authenticated-orcid":false,"given":"Yangyang","family":"Xu","sequence":"additional","affiliation":[{"name":"Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei, Anhui, China and University of Science and Technology of China, Hefei, Anhui, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-5847-5206","authenticated-orcid":false,"given":"Qitao","family":"Qin","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, Anhui, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8253-1785","authenticated-orcid":false,"given":"Xianwei","family":"Meng","sequence":"additional","affiliation":[{"name":"Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei, Anhui, China"}]}],"member":"320","published-online":{"date-parts":[[2025,8,3]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01499"},{"key":"e_1_3_2_2_2_1","volume-title":"Revisiting feature prediction for learning visual representations from video. arXiv preprint arXiv:2404.08471","author":"Bardes Adrien","year":"2024","unstructured":"Adrien Bardes, Quentin Garrido, Jean Ponce, Xinlei Chen, Michael Rabbat, Yann LeCun, Mahmoud Assran, and Nicolas Ballas. 2024. Revisiting feature prediction for learning visual representations from video. arXiv preprint arXiv:2404.08471 (2024)."},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1080\/01621459.1970.10481180"},{"key":"e_1_3_2_2_4_1","volume-title":"Time-series analysis. Intelligent data analysis: An introduction","author":"Bradley Elizabeth","year":"1999","unstructured":"Elizabeth Bradley. 1999. Time-series analysis. Intelligent data analysis: An introduction (1999), 167-194."},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i6.25854"},{"key":"e_1_3_2_2_6_1","volume-title":"Tsmixer: An all-mlp architecture for time series forecasting. arXiv preprint arXiv:2303.06053","author":"Chen Si-An","year":"2023","unstructured":"Si-An Chen, Chun-Liang Li, Nate Yoder, Sercan O Arik, and Tomas Pfister. 2023. Tsmixer: An all-mlp architecture for time series forecasting. arXiv preprint arXiv:2303.06053 (2023)."},{"key":"e_1_3_2_2_7_1","volume-title":"Long-term forecasting with tide: Time-series dense encoder. arXiv preprint arXiv:2304.08424","author":"Das Abhimanyu","year":"2023","unstructured":"Abhimanyu Das, Weihao Kong, Andrew Leach, Shaan Mathur, Rajat Sen, and Rose Yu. 2023. Long-term forecasting with tide: Time-series dense encoder. arXiv preprint arXiv:2304.08424 (2023)."},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1214\/06-BA131"},{"key":"e_1_3_2_2_9_1","volume-title":"Time-Series JEPA for Predictive Remote Control under Capacity-Limited Networks. arXiv preprint arXiv:2406.04853","author":"Girgis Abanoub M","year":"2024","unstructured":"Abanoub M Girgis, Alvaro Valcarce, and Mehdi Bennis. 2024. Time-Series JEPA for Predictive Remote Control under Capacity-Limited Networks. arXiv preprint arXiv:2406.04853 (2024)."},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00778-020-00608-7"},{"key":"e_1_3_2_2_11_1","volume-title":"Comparison of ARIMA and Random Forest time series models for prediction of avian influenza H5N1 outbreaks. BMC bioinformatics","author":"Kane Michael J","year":"2014","unstructured":"Michael J Kane, Natalie Price, Matthew Scotch, and Peter Rabinowitz. 2014. Comparison of ARIMA and Random Forest time series models for prediction of avian influenza H5N1 outbreaks. BMC bioinformatics, Vol. 15 (2014), 1-9."},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2019.12.030"},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.laa.2015.07.021"},{"key":"e_1_3_2_2_14_1","volume-title":"Supervised contrastive learning. Advances in neural information processing systems","author":"Khosla Prannay","year":"2020","unstructured":"Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. 2020. Supervised contrastive learning. Advances in neural information processing systems, Vol. 33 (2020), 18661-18673."},{"key":"e_1_3_2_2_15_1","volume-title":"International Conference on Learning Representations.","author":"Kim Taesung","year":"2021","unstructured":"Taesung Kim, Jinhee Kim, Yunwon Tae, Cheonbok Park, Jang-Ho Choi, and Jaegul Choo. 2021. Reversible instance normalization for accurate time-series forecasting against distribution shift. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3210006"},{"key":"e_1_3_2_2_17_1","first-page":"1","article-title":"A path towards autonomous machine intelligence version 0.9. 2, 2022-06-27","volume":"62","author":"LeCun Yann","year":"2022","unstructured":"Yann LeCun. 2022. A path towards autonomous machine intelligence version 0.9. 2, 2022-06-27. Open Review, Vol. 62, 1 (2022), 1-62.","journal-title":"Open Review"},{"key":"e_1_3_2_2_18_1","first-page":"5816","article-title":"Scinet: Time series modeling and forecasting with sample convolution and interaction","volume":"35","author":"Liu Minhao","year":"2022","unstructured":"Minhao Liu, Ailing Zeng, Muxi Chen, Zhijian Xu, Qiuxia Lai, Lingna Ma, and Qiang Xu. 2022b. Scinet: Time series modeling and forecasting with sample convolution and interaction. Advances in Neural Information Processing Systems, Vol. 35 (2022), 5816-5828.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_19_1","volume-title":"International conference on learning representations.","author":"Liu Shizhan","year":"2021","unstructured":"Shizhan Liu, Hang Yu, Cong Liao, Jianguo Li, Weiyao Lin, Alex X Liu, and Schahram Dustdar. 2021. Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting. In International conference on learning representations."},{"key":"e_1_3_2_2_20_1","volume-title":"itransformer: Inverted transformers are effective for time series forecasting. arXiv preprint arXiv:2310.06625","author":"Liu Yong","year":"2024","unstructured":"Yong Liu, Tengge Hu, Haoran Zhang, Haixu Wu, Shiyu Wang, Lintao Ma, and Mingsheng Long. 2024b. itransformer: Inverted transformers are effective for time series forecasting. arXiv preprint arXiv:2310.06625 (2024)."},{"key":"e_1_3_2_2_21_1","first-page":"9881","article-title":"Non-stationary transformers: Exploring the stationarity in time series forecasting","volume":"35","author":"Liu Yong","year":"2022","unstructured":"Yong Liu, Haixu Wu, Jianmin Wang, and Mingsheng Long. 2022a. Non-stationary transformers: Exploring the stationarity in time series forecasting. Advances in Neural Information Processing Systems, Vol. 35 (2022), 9881-9893.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_22_1","volume-title":"Advances in Neural Information Processing Systems","volume":"36","author":"Liu Zhiding","year":"2024","unstructured":"Zhiding Liu, Mingyue Cheng, Zhi Li, Zhenya Huang, Qi Liu, Yanhu Xie, and Enhong Chen. 2024a. Adaptive normalization for non-stationary time series forecasting: A temporal slice perspective. Advances in Neural Information Processing Systems, Vol. 36 (2024)."},{"key":"e_1_3_2_2_23_1","volume-title":"Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101","author":"Loshchilov I","year":"2017","unstructured":"I Loshchilov. 2017. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)."},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00522"},{"key":"e_1_3_2_2_25_1","volume-title":"The Eleventh International Conference on Learning Representations.","author":"Nie Yuqi","year":"2023","unstructured":"Yuqi Nie, Nam H Nguyen, Phanwadee Sinthong, and Jayant Kalagnanam. 2023. A Time Series is Worth 64 Words: Long-term Forecasting with Transformers. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_2_26_1","volume-title":"N-BEATS: Neural basis expansion analysis for interpretable time series forecasting. arXiv preprint arXiv:1905.10437","author":"Oreshkin Boris N","year":"2019","unstructured":"Boris N Oreshkin, Dmitri Carpov, Nicolas Chapados, and Yoshua Bengio. 2019. N-BEATS: Neural basis expansion analysis for interpretable time series forecasting. arXiv preprint arXiv:1905.10437 (2019)."},{"key":"e_1_3_2_2_27_1","volume-title":"Tfb: Towards comprehensive and fair benchmarking of time series forecasting methods. arXiv preprint arXiv:2403.20150","author":"Qiu Xiangfei","year":"2024","unstructured":"Xiangfei Qiu, Jilin Hu, Lekui Zhou, Xingjian Wu, Junyang Du, Buang Zhang, Chenjuan Guo, Aoying Zhou, Christian S Jensen, Zhenli Sheng, et al., 2024. Tfb: Towards comprehensive and fair benchmarking of time series forecasting methods. arXiv preprint arXiv:2403.20150 (2024)."},{"key":"e_1_3_2_2_28_1","volume-title":"Mehmet Ugur Gudelek, and Ahmet Murat Ozbayoglu","author":"Sezer Omer Berat","year":"2020","unstructured":"Omer Berat Sezer, Mehmet Ugur Gudelek, and Ahmet Murat Ozbayoglu. 2020. Financial time series forecasting with deep learning: A systematic literature review: 2005-2019. Applied soft computing, Vol. 90 (2020), 106181."},{"key":"e_1_3_2_2_29_1","unstructured":"Zezhi Shao Fei Wang Yongjun Xu Wei Wei Chengqing Yu Zhao Zhang Di Yao Tao Sun Guangyin Jin Xin Cao et al. 2024. Exploring progress in multivariate time series forecasting: Comprehensive benchmarking and heterogeneity analysis. IEEE Transactions on Knowledge and Data Engineering (2024)."},{"key":"e_1_3_2_2_30_1","volume-title":"Vector autoregression and causality: a theoretical overview and simulation study. Econometric reviews","author":"Toda Hiro Y","year":"1994","unstructured":"Hiro Y Toda and Peter CB Phillips. 1994. Vector autoregression and causality: a theoretical overview and simulation study. Econometric reviews, Vol. 13, 2 (1994), 259-285."},{"key":"e_1_3_2_2_31_1","volume-title":"Practical lossless compression with latent variables using bits back coding. arXiv preprint arXiv:1901.04866","author":"Townsend James","year":"2019","unstructured":"James Townsend, Tom Bird, and David Barber. 2019. Practical lossless compression with latent variables using bits back coding. arXiv preprint arXiv:1901.04866 (2019)."},{"key":"e_1_3_2_2_32_1","volume-title":"Attention is all you need. Advances in Neural Information Processing Systems","author":"Vaswani A","year":"2017","unstructured":"A Vaswani. 2017. Attention is all you need. Advances in Neural Information Processing Systems (2017)."},{"key":"e_1_3_2_2_33_1","volume-title":"LaT-PFN: A Joint Embedding Predictive Architecture for In-context Time-series Forecasting. arXiv preprint arXiv:2405.10093","author":"Verdenius Stijn","year":"2024","unstructured":"Stijn Verdenius, Andrea Zerio, and Roy LM Wang. 2024. LaT-PFN: A Joint Embedding Predictive Architecture for In-context Time-series Forecasting. arXiv preprint arXiv:2405.10093 (2024)."},{"key":"e_1_3_2_2_34_1","volume-title":"Timemixer: Decomposable multiscale mixing for time series forecasting. arXiv preprint arXiv:2405.14616","author":"Wang Shiyu","year":"2024","unstructured":"Shiyu Wang, Haixu Wu, Xiaoming Shi, Tengge Hu, Huakun Luo, Lintao Ma, James Y Zhang, and Jun Zhou. 2024. Timemixer: Decomposable multiscale mixing for time series forecasting. arXiv preprint arXiv:2405.14616 (2024)."},{"key":"e_1_3_2_2_35_1","volume-title":"International Conference on Machine Learning. PMLR, 22964-22984","author":"Wang Thomas","year":"2022","unstructured":"Thomas Wang, Adam Roberts, Daniel Hesslow, Teven Le Scao, Hyung Won Chung, Iz Beltagy, Julien Launay, and Colin Raffel. 2022. What language model architecture and pretraining objective works best for zero-shot generalization?. In International Conference on Machine Learning. PMLR, 22964-22984."},{"key":"e_1_3_2_2_36_1","volume-title":"Transformers in time series: A survey. arXiv preprint arXiv:2202.07125","author":"Wen Qingsong","year":"2022","unstructured":"Qingsong Wen, Tian Zhou, Chaoli Zhang, Weiqi Chen, Ziqing Ma, Junchi Yan, and Liang Sun. 2022. Transformers in time series: A survey. arXiv preprint arXiv:2202.07125 (2022)."},{"key":"e_1_3_2_2_37_1","volume-title":"Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv preprint arXiv:2210.02186","author":"Wu Haixu","year":"2022","unstructured":"Haixu Wu, Tengge Hu, Yong Liu, Hang Zhou, Jianmin Wang, and Mingsheng Long. 2022. Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv preprint arXiv:2210.02186 (2022)."},{"key":"e_1_3_2_2_38_1","volume-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in neural information processing systems","author":"Wu Haixu","year":"2021","unstructured":"Haixu Wu, Jiehui Xu, Jianmin Wang, and Mingsheng Long. 2021. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in neural information processing systems, Vol. 34 (2021), 22419-22430."},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00060"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"e_1_3_2_2_41_1","volume-title":"Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In The eleventh international conference on learning representations.","author":"Zhang Yunhao","year":"2023","unstructured":"Yunhao Zhang and Junchi Yan. 2023. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In The eleventh international conference on learning representations."},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"e_1_3_2_2_43_1","volume-title":"Film: Frequency improved legendre memory model for long-term time series forecasting. Advances in neural information processing systems","author":"Zhou Tian","year":"2022","unstructured":"Tian Zhou, Ziqing Ma, Qingsong Wen, Liang Sun, Tao Yao, Wotao Yin, Rong Jin, et al., 2022a. Film: Frequency improved legendre memory model for long-term time series forecasting. Advances in neural information processing systems, Vol. 35 (2022), 12677-12690."},{"key":"e_1_3_2_2_44_1","volume-title":"International conference on machine learning. PMLR, 27268-27286","author":"Zhou Tian","year":"2022","unstructured":"Tian Zhou, Ziqing Ma, Qingsong Wen, Xue Wang, Liang Sun, and Rong Jin. 2022b. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In International conference on machine learning. PMLR, 27268-27286."}],"event":{"name":"KDD '25: The 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Toronto ON Canada","acronym":"KDD '25","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V.2"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3711896.3737157","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,16]],"date-time":"2025-08-16T14:41:51Z","timestamp":1755355311000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3711896.3737157"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,3]]},"references-count":44,"alternative-id":["10.1145\/3711896.3737157","10.1145\/3711896"],"URL":"https:\/\/doi.org\/10.1145\/3711896.3737157","relation":{},"subject":[],"published":{"date-parts":[[2025,8,3]]},"assertion":[{"value":"2025-08-03","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}