{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T13:11:21Z","timestamp":1775913081717,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":48,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,5,8]]},"DOI":"10.1145\/3701716.3715214","type":"proceedings-article","created":{"date-parts":[[2025,6,23]],"date-time":"2025-06-23T14:10:32Z","timestamp":1750687832000},"page":"171-180","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":27,"title":["ConvTimeNet: A Deep Hierarchical Fully Convolutional Model for Multivariate Time Series Analysis"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9873-7681","authenticated-orcid":false,"given":"Mingyue","family":"Cheng","sequence":"first","affiliation":[{"name":"State Key Laboratory of Cognitive Intelligence, University of Science and Technology of China, Hefei, Anhui Province, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-6421-2626","authenticated-orcid":false,"given":"Jiqian","family":"Yang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Cognitive Intelligence, University of Science and Technology of China, Hefei, Anhui Province, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-4267-9111","authenticated-orcid":false,"given":"Tingyue","family":"Pan","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Cognitive Intelligence, University of Science and Technology of China, Hefei, Anhui Province, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6956-5550","authenticated-orcid":false,"given":"Qi","family":"Liu","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Cognitive Intelligence, University of Science and Technology of China, Hefei, Anhui Province, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8061-7486","authenticated-orcid":false,"given":"Zhi","family":"Li","sequence":"additional","affiliation":[{"name":"Shenzhen International Graduate School, Tsinghua University, Shenzhen, Guangdong Province, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9202-7678","authenticated-orcid":false,"given":"Shijin","family":"Wang","sequence":"additional","affiliation":[{"name":"Artificial Intelligence Research Institute, iFLYTEK Co., Ltd, Hefei, Anhui Province, China"}]}],"member":"320","published-online":{"date-parts":[[2025,5,23]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"Henry Mao, Gary Cottrell, and Julian McAuley.","author":"Bachlechner Thomas","year":"2021","unstructured":"Thomas Bachlechner, Bodhisattwa Prasad Majumder, Henry Mao, Gary Cottrell, and Julian McAuley. 2021. Rezero is all you need: Fast convergence at large depth. In Uncertainty in Artificial Intelligence. PMLR, 1352--1361."},{"key":"e_1_3_2_2_2_1","volume-title":"Aaron Bostrom, James Large, and Eamonn Keogh.","author":"Bagnall Anthony","year":"2017","unstructured":"Anthony Bagnall, Jason Lines, Aaron Bostrom, James Large, and Eamonn Keogh. 2017. The great time series classification bake off: a review and experimental evaluation of recent algorithmic advances. Data mining and knowledge discovery, Vol. 31 (2017), 606--660."},{"key":"e_1_3_2_2_3_1","volume-title":"An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv:1803.01271","author":"Bai Shaojie","year":"2018","unstructured":"Shaojie Bai, J Zico Kolter, and Vladlen Koltun. 2018. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv:1803.01271 (2018)."},{"key":"e_1_3_2_2_4_1","volume-title":"Time series analysis: forecasting and control","author":"Box George EP","unstructured":"George EP Box, Gwilym M Jenkins, Gregory C Reinsel, and Greta M Ljung. 2015. Time series analysis: forecasting and control. John Wiley & Sons."},{"key":"e_1_3_2_2_5_1","volume-title":"MSGNet: Learning Multi-Scale Inter-Series Correlations for Multivariate Time Series Forecasting. arXiv preprint arXiv:2401.00423","author":"Cai Wanlin","year":"2023","unstructured":"Wanlin Cai, Yuxuan Liang, Xianggen Liu, Jianshuai Feng, and Yuankai Wu. 2023. MSGNet: Learning Multi-Scale Inter-Series Correlations for Multivariate Time Series Forecasting. arXiv preprint arXiv:2401.00423 (2023)."},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i6.25854"},{"key":"e_1_3_2_2_7_1","volume-title":"Llm4ts: Two-stage fine-tuning for time-series forecasting with pre-trained llms. arXiv preprint arXiv:2308.08469","author":"Chang Ching","year":"2023","unstructured":"Ching Chang, Wen-Chih Peng, and Tien-Fu Chen. 2023. Llm4ts: Two-stage fine-tuning for time-series forecasting with pre-trained llms. arXiv preprint arXiv:2308.08469 (2023)."},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475467"},{"key":"e_1_3_2_2_9_1","volume-title":"Advancing Time Series Classification with Multimodal Language Modeling. arXiv preprint arXiv:2403.12371","author":"Cheng Mingyue","year":"2024","unstructured":"Mingyue Cheng, Yiheng Chen, Qi Liu, Zhiding Liu, and Yucong Luo. 2024. Advancing Time Series Classification with Multimodal Language Modeling. arXiv preprint arXiv:2403.12371 (2024)."},{"key":"e_1_3_2_2_10_1","volume-title":"FormerTime: Hierarchical Multi-Scale Representations for Multivariate Time Series Classification. arXiv preprint arXiv:2302.09818","author":"Cheng Mingyue","year":"2023","unstructured":"Mingyue Cheng, Qi Liu, Zhiding Liu, Zhi Li, Yucong Luo, and Enhong Chen. 2023a. FormerTime: Hierarchical Multi-Scale Representations for Multivariate Time Series Classification. arXiv preprint arXiv:2302.09818 (2023)."},{"key":"e_1_3_2_2_11_1","volume-title":"TimeMAE: Self-Supervised Representations of Time Series with Decoupled Masked Autoencoders. arXiv preprint arXiv:2303.00320","author":"Cheng Mingyue","year":"2023","unstructured":"Mingyue Cheng, Qi Liu, Zhiding Liu, Hao Zhang, Rujiao Zhang, and Enhong Chen. 2023b. TimeMAE: Self-Supervised Representations of Time Series with Decoupled Masked Autoencoders. arXiv preprint arXiv:2303.00320 (2023)."},{"key":"e_1_3_2_2_12_1","volume-title":"Multi-scale convolutional neural networks for time series classification. arXiv preprint arXiv:1603.06995","author":"Cui Zhicheng","year":"2016","unstructured":"Zhicheng Cui, Wenlin Chen, and Yixin Chen. 2016. Multi-scale convolutional neural networks for time series classification. arXiv preprint arXiv:1603.06995 (2016)."},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.89"},{"key":"e_1_3_2_2_14_1","volume-title":"Long-term Forecasting with","author":"Das Abhimanyu","year":"2023","unstructured":"Abhimanyu Das, Weihao Kong, Andrew Leach, Rajat Sen, and Rose Yu. 2023. Long-term Forecasting with TiDE: Time-series Dense Encoder. arXiv preprint arXiv:2304.08424 (2023)."},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467231"},{"key":"e_1_3_2_2_16_1","volume-title":"Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. 11963--11975","author":"Ding Xiaohan","year":"2022","unstructured":"Xiaohan Ding, Xiangyu Zhang, Jungong Han, and Guiguang Ding. 2022. Scaling up your kernels to 31x31: Revisiting large kernel design in cnns. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. 11963--11975."},{"key":"e_1_3_2_2_17_1","unstructured":"Alexey Dosovitskiy Lucas Beyer Alexander Kolesnikov Dirk Weissenborn Xiaohua Zhai Thomas Unterthiner Mostafa Dehghani Matthias Minderer Georg Heigold Sylvain Gelly et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)."},{"key":"e_1_3_2_2_18_1","volume-title":"Mamba: Linear-time sequence modeling with selective state spaces. arXiv preprint arXiv:2312.00752","author":"Gu Albert","year":"2023","unstructured":"Albert Gu and Tri Dao. 2023. Mamba: Linear-time sequence modeling with selective state spaces. arXiv preprint arXiv:2312.00752 (2023)."},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_2_20_1","volume-title":"ShapeWordNet: An Interpretable Shapelet Neural Network for Physiological Signal Classification. In International Conference on Database Systems for Advanced Applications. Springer, 353--369","author":"He Wenqiang","year":"2023","unstructured":"Wenqiang He, Mingyue Cheng, Qi Liu, and Zhi Li. 2023. ShapeWordNet: An Interpretable Shapelet Neural Network for Physiological Signal Classification. In International Conference on Database Systems for Advanced Applications. Springer, 353--369."},{"key":"e_1_3_2_2_21_1","volume-title":"International conference on machine learning. pmlr, 448--456","author":"Ioffe Sergey","year":"2015","unstructured":"Sergey Ioffe and Christian Szegedy. 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International conference on machine learning. pmlr, 448--456."},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10618-020-00710-y"},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i7.25976"},{"key":"e_1_3_2_2_24_1","volume-title":"Time-llm: Time series forecasting by reprogramming large language models. arXiv preprint arXiv:2310.01728","author":"Jin Ming","year":"2023","unstructured":"Ming Jin, Shiyu Wang, Lintao Ma, Zhixuan Chu, James Y Zhang, Xiaoming Shi, Pin-Yu Chen, Yuxuan Liang, Yuan-Fang Li, Shirui Pan, et al. 2023. Time-llm: Time series forecasting by reprogramming large language models. arXiv preprint arXiv:2310.01728 (2023)."},{"key":"e_1_3_2_2_25_1","volume-title":"International Conference on Learning Representations.","author":"Kim Taesung","year":"2021","unstructured":"Taesung Kim, Jinhee Kim, Yunwon Tae, Cheonbok Park, Jang-Ho Choi, and Jaegul Choo. 2021. Reversible instance normalization for accurate time-series forecasting against distribution shift. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_26_1","volume-title":"Gated transformer networks for multivariate time series classification. arXiv preprint arXiv:2103.14438","author":"Liu Minghao","year":"2021","unstructured":"Minghao Liu, Shengqi Ren, Siyuan Ma, Jiahui Jiao, Yizhou Chen, Zhiguang Wang, and Wei Song. 2021a. Gated transformer networks for multivariate time series classification. arXiv preprint arXiv:2103.14438 (2021)."},{"key":"e_1_3_2_2_27_1","volume-title":"itransformer: Inverted transformers are effective for time series forecasting. arXiv preprint arXiv:2310.06625","author":"Liu Yong","year":"2023","unstructured":"Yong Liu, Tengge Hu, Haoran Zhang, Haixu Wu, Shiyu Wang, Lintao Ma, and Mingsheng Long. 2023. itransformer: Inverted transformers are effective for time series forecasting. arXiv preprint arXiv:2310.06625 (2023)."},{"key":"e_1_3_2_2_28_1","volume-title":"Generative Pretrained Hierarchical Transformer for Time Series Forecasting. arXiv preprint arXiv:2402.16516","author":"Liu Zhiding","year":"2024","unstructured":"Zhiding Liu, Jiqian Yang, Mingyue Cheng, Yucong Luo, and Zhi Li. 2024. Generative Pretrained Hierarchical Transformer for Time Series Forecasting. arXiv preprint arXiv:2402.16516 (2024)."},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3091162"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.sysarc.2022.102443"},{"key":"e_1_3_2_2_31_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10994-021-06057-9"},{"key":"e_1_3_2_2_32_1","volume-title":"Bake off redux: a review and experimental evaluation of recent time series classification algorithms. arXiv preprint arXiv:2304.13029","author":"Middlehurst Matthew","year":"2023","unstructured":"Matthew Middlehurst, Patrick Sch\u00e4fer, and Anthony Bagnall. 2023. Bake off redux: a review and experimental evaluation of recent time series classification algorithms. arXiv preprint arXiv:2304.13029 (2023)."},{"key":"e_1_3_2_2_33_1","volume-title":"A time series is worth 64 words: Long-term forecasting with transformers. arXiv preprint arXiv:2211.14730","author":"Nie Yuqi","year":"2022","unstructured":"Yuqi Nie, Nam H Nguyen, Phanwadee Sinthong, and Jayant Kalagnanam. 2022. A time series is worth 64 words: Long-term forecasting with transformers. arXiv preprint arXiv:2211.14730 (2022)."},{"key":"e_1_3_2_2_34_1","volume-title":"Recurrent neural networks for time series forecasting. arXiv preprint arXiv:1901.00069","author":"Petneh\u00e1zi G\u00e1bor","year":"2019","unstructured":"G\u00e1bor Petneh\u00e1zi. 2019. Recurrent neural networks for time series forecasting. arXiv preprint arXiv:1901.00069 (2019)."},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1098\/rsta.2011.0550"},{"key":"e_1_3_2_2_36_1","volume-title":"International conference on machine learning. PMLR, 6105--6114","author":"Tan Mingxing","year":"2019","unstructured":"Mingxing Tan and Quoc Le. 2019. Efficientnet: Rethinking model scaling for convolutional neural networks. In International conference on machine learning. PMLR, 6105--6114."},{"key":"e_1_3_2_2_37_1","volume-title":"Are language models actually useful for time series forecasting? arXiv preprint arXiv:2406.16964","author":"Tan Mingtian","year":"2024","unstructured":"Mingtian Tan, Mike A Merrill, Vinayak Gupta, Tim Althoff, and Thomas Hartvigsen. 2024. Are language models actually useful for time series forecasting? arXiv preprint arXiv:2406.16964 (2024)."},{"key":"e_1_3_2_2_38_1","volume-title":"Attention is all you need. Advances in neural information processing systems","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems, Vol. 30 (2017)."},{"key":"e_1_3_2_2_39_1","volume-title":"Transformers in time series: A survey. arXiv preprint arXiv:2202.07125","author":"Wen Qingsong","year":"2022","unstructured":"Qingsong Wen, Tian Zhou, Chaoli Zhang, Weiqi Chen, Ziqing Ma, Junchi Yan, and Liang Sun. 2022. Transformers in time series: A survey. arXiv preprint arXiv:2202.07125 (2022)."},{"key":"e_1_3_2_2_40_1","volume-title":"Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv preprint arXiv:2210.02186","author":"Wu Haixu","year":"2022","unstructured":"Haixu Wu, Tengge Hu, Yong Liu, Hang Zhou, Jianmin Wang, and Mingsheng Long. 2022. Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv preprint arXiv:2210.02186 (2022)."},{"key":"e_1_3_2_2_41_1","first-page":"22419","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","volume":"34","author":"Wu Haixu","year":"2021","unstructured":"Haixu Wu, Jiehui Xu, Jianmin Wang, and Mingsheng Long. 2021. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in Neural Information Processing Systems, Vol. 34 (2021), 22419--22430.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/1557019.1557122"},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467401"},{"key":"e_1_3_2_2_45_1","doi-asserted-by":"publisher","DOI":"10.1016\/S0925-2312(01)00702-0"},{"key":"e_1_3_2_2_46_1","volume-title":"The Eleventh International Conference on Learning Representations.","author":"Zhang Yunhao","year":"2022","unstructured":"Yunhao Zhang and Junchi Yan. 2022. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_2_47_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-08010-9_33"},{"key":"e_1_3_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"}],"event":{"name":"WWW '25: The ACM Web Conference 2025","location":"Sydney NSW Australia","acronym":"WWW '25","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Companion Proceedings of the ACM on Web Conference 2025"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3701716.3715214","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,7]],"date-time":"2025-10-07T18:24:47Z","timestamp":1759861487000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3701716.3715214"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,8]]},"references-count":48,"alternative-id":["10.1145\/3701716.3715214","10.1145\/3701716"],"URL":"https:\/\/doi.org\/10.1145\/3701716.3715214","relation":{},"subject":[],"published":{"date-parts":[[2025,5,8]]},"assertion":[{"value":"2025-05-23","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}