{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,17]],"date-time":"2026-04-17T15:59:47Z","timestamp":1776441587326,"version":"3.51.2"},"publisher-location":"New York, NY, USA","reference-count":50,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,10,21]],"date-time":"2023-10-21T00:00:00Z","timestamp":1697846400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100004739","name":"Youth Innovation Promotion Association of the Chinese Academy of Sciences","doi-asserted-by":"publisher","award":["No.2023112"],"award-info":[{"award-number":["No.2023112"]}],"id":[{"id":"10.13039\/501100004739","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,10,21]]},"DOI":"10.1145\/3583780.3614851","type":"proceedings-article","created":{"date-parts":[[2023,10,21]],"date-time":"2023-10-21T07:45:26Z","timestamp":1697874326000},"page":"3062-3072","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":97,"title":["DSformer: A Double Sampling Transformer for Multivariate Time Series Long-term Prediction"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8314-8251","authenticated-orcid":false,"given":"Chengqing","family":"Yu","sequence":"first","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences &amp; University of Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3282-0535","authenticated-orcid":false,"given":"Fei","family":"Wang","sequence":"additional","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences &amp; University of Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0815-2768","authenticated-orcid":false,"given":"Zezhi","family":"Shao","sequence":"additional","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences &amp; University of Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1692-3574","authenticated-orcid":false,"given":"Tao","family":"Sun","sequence":"additional","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0574-2479","authenticated-orcid":false,"given":"Lin","family":"Wu","sequence":"additional","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6647-0986","authenticated-orcid":false,"given":"Yongjun","family":"Xu","sequence":"additional","affiliation":[{"name":"Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2023,10,21]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2021.12.001"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i6.20587"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICOIN53446.2022.9687205"},{"key":"e_1_3_2_1_4_1","volume-title":"Triformer: Triangular, Variable-Specific Attentions for Long Sequence Multivariate Time Series Forecasting--Full Version. arXiv preprint arXiv:2204.13767","author":"Cirstea Razvan-Gabriel","year":"2022","unstructured":"Razvan-Gabriel Cirstea , Chenjuan Guo , Bin Yang , Tung Kieu , Xuanyi Dong , and Shirui Pan . 2022 . Triformer: Triangular, Variable-Specific Attentions for Long Sequence Multivariate Time Series Forecasting--Full Version. arXiv preprint arXiv:2204.13767 (2022). Razvan-Gabriel Cirstea, Chenjuan Guo, Bin Yang, Tung Kieu, Xuanyi Dong, and Shirui Pan. 2022. Triformer: Triangular, Variable-Specific Attentions for Long Sequence Multivariate Time Series Forecasting--Full Version. arXiv preprint arXiv:2204.13767 (2022)."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3456389.3456397"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i6.20618"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i6.20618"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583312"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2018.2866149"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2022.105560"},{"key":"e_1_3_2_1_11_1","volume-title":"Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980","author":"Kingma Diederik P","year":"2014","unstructured":"Diederik P Kingma and Jimmy Ba . 2014 . Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014). Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)."},{"key":"e_1_3_2_1_12_1","volume-title":"Reformer: The Efficient Transformer. In International Conference on Learning Representations.","author":"Kitaev Nikita","year":"2019","unstructured":"Nikita Kitaev , Lukasz Kaiser , and Anselm Levskaya . 2019 . Reformer: The Efficient Transformer. In International Conference on Learning Representations. Nikita Kitaev, Lukasz Kaiser, and Anselm Levskaya. 2019. Reformer: The Efficient Transformer. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.envpol.2022.119348"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3210006"},{"key":"e_1_3_2_1_15_1","volume-title":"Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Advances in neural information processing systems","author":"Li Shiyang","year":"2019","unstructured":"Shiyang Li , Xiaoyong Jin , Yao Xuan , Xiyou Zhou , Wenhu Chen , Yu-Xiang Wang , and Xifeng Yan . 2019. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Advances in neural information processing systems , Vol. 32 ( 2019 ). Shiyang Li, Xiaoyong Jin, Yao Xuan, Xiyou Zhou, Wenhu Chen, Yu-Xiang Wang, and Xifeng Yan. 2019. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Advances in neural information processing systems, Vol. 32 (2019)."},{"key":"e_1_3_2_1_16_1","volume-title":"2023 a. Knowledge Graph Contrastive Learning Based on Relation-Symmetrical Structure","author":"Liang Ke","year":"2023","unstructured":"Ke Liang , Yue Liu , Sihang Zhou , Wenxuan Tu , Yi Wen , Xihong Yang , Xiangjun Dong , and Xinwang Liu . 2023 a. Knowledge Graph Contrastive Learning Based on Relation-Symmetrical Structure . IEEE Transactions on Knowledge and Data Engineering ( 2023 ). Ke Liang, Yue Liu, Sihang Zhou, Wenxuan Tu, Yi Wen, Xihong Yang, Xiangjun Dong, and Xinwang Liu. 2023 a. Knowledge Graph Contrastive Learning Based on Relation-Symmetrical Structure. IEEE Transactions on Knowledge and Data Engineering (2023)."},{"key":"e_1_3_2_1_17_1","volume-title":"2023 b. Message Intercommunication for Inductive Relation Reasoning. arXiv preprint arXiv:2305.14074","author":"Liang Ke","year":"2023","unstructured":"Ke Liang , Lingyuan Meng , Sihang Zhou , Siwei Wang , Wenxuan Tu , Yue Liu , Meng Liu , and Xinwang Liu . 2023 b. Message Intercommunication for Inductive Relation Reasoning. arXiv preprint arXiv:2305.14074 ( 2023 ). Ke Liang, Lingyuan Meng, Sihang Zhou, Siwei Wang, Wenxuan Tu, Yue Liu, Meng Liu, and Xinwang Liu. 2023 b. Message Intercommunication for Inductive Relation Reasoning. arXiv preprint arXiv:2305.14074 (2023)."},{"key":"e_1_3_2_1_18_1","volume-title":"Benchmarking, Measuring, and Optimizing: 14th BenchCouncil International Symposium, Bench","author":"Liang Yubo","year":"2022","unstructured":"Yubo Liang , Zezhi Shao , Fei Wang , Zhao Zhang , Tao Sun , and Yongjun Xu . 2023 c. BasicTS: An Open Source Fair Multivariate Time Series Prediction Benchmark . In Benchmarking, Measuring, and Optimizing: 14th BenchCouncil International Symposium, Bench 2022 , Virtual Event, November 7--9, 2022, Revised Selected Papers. Springer , 87--101. Yubo Liang, Zezhi Shao, Fei Wang, Zhao Zhang, Tao Sun, and Yongjun Xu. 2023 c. BasicTS: An Open Source Fair Multivariate Time Series Prediction Benchmark. In Benchmarking, Measuring, and Optimizing: 14th BenchCouncil International Symposium, Bench 2022, Virtual Event, November 7--9, 2022, Revised Selected Papers. Springer, 87--101."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.energy.2020.117794"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.measurement.2021.109347"},{"key":"e_1_3_2_1_21_1","volume-title":"International Conference on Learning Representations.","author":"Liu Shizhan","year":"2022","unstructured":"Shizhan Liu , Hang Yu , Cong Liao , Jianguo Li , Weiyao Lin , Alex X Liu , and Schahram Dustdar . 2022 b. Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting . In International Conference on Learning Representations. Shizhan Liu, Hang Yu, Cong Liao, Jianguo Li, Weiyao Lin, Alex X Liu, and Schahram Dustdar. 2022b. Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.apr.2021.101197"},{"key":"e_1_3_2_1_23_1","unstructured":"Yijing Liu Qinxian Liu Jian-Wei Zhang Haozhe Feng Zhongwei Wang Zihan Zhou and Wei Chen. 2022a. Multivariate Time-Series Forecasting with Temporal Polynomial Graph Neural Networks. In Advances in Neural Information Processing Systems.  Yijing Liu Qinxian Liu Jian-Wei Zhang Haozhe Feng Zhongwei Wang Zihan Zhou and Wei Chen. 2022a. Multivariate Time-Series Forecasting with Temporal Polynomial Graph Neural Networks. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599838"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2022.103643"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3538712.3538723"},{"key":"e_1_3_2_1_27_1","volume-title":"The Eleventh International Conference on Learning Representations.","author":"Nie Yuqi","year":"2023","unstructured":"Yuqi Nie , Nam H Nguyen , Phanwadee Sinthong , and Jayant Kalagnanam . 2023 . A Time Series is Worth 64 Words: Long-term Forecasting with Transformers . In The Eleventh International Conference on Learning Representations. Yuqi Nie, Nam H Nguyen, Phanwadee Sinthong, and Jayant Kalagnanam. 2023. A Time Series is Worth 64 Words: Long-term Forecasting with Transformers. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_28_1","volume-title":"HUTFormer: Hierarchical U-Net Transformer for Long-Term Traffic Forecasting. arXiv preprint arXiv:2307.14596","author":"Shao Zezhi","year":"2023","unstructured":"Zezhi Shao , Fei Wang , Zhao Zhang , Yuchen Fang , Guangyin Jin , and Yongjun Xu. 2023. HUTFormer: Hierarchical U-Net Transformer for Long-Term Traffic Forecasting. arXiv preprint arXiv:2307.14596 ( 2023 ). Zezhi Shao, Fei Wang, Zhao Zhang, Yuchen Fang, Guangyin Jin, and Yongjun Xu. 2023. HUTFormer: Hierarchical U-Net Transformer for Long-Term Traffic Forecasting. arXiv preprint arXiv:2307.14596 (2023)."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3511808.3557702"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539396"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.14778\/3551793.3551827"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chaos.2022.112405"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.xinn.2023.100405"},{"key":"e_1_3_2_1_34_1","volume-title":"The Eleventh International Conference on Learning Representations.","author":"Wu Haixu","year":"2023","unstructured":"Haixu Wu , Tengge Hu , Yong Liu , Hang Zhou , Jianmin Wang , and Mingsheng Long . 2023 . TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis . In The Eleventh International Conference on Learning Representations. Haixu Wu, Tengge Hu, Yong Liu, Hang Zhou, Jianmin Wang, and Mingsheng Long. 2023. TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_35_1","first-page":"22419","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","volume":"34","author":"Wu Haixu","year":"2021","unstructured":"Haixu Wu , Jiehui Xu , Jianmin Wang , and Mingsheng Long . 2021 a. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting . Advances in Neural Information Processing Systems , Vol. 34 (2021), 22419 -- 22430 . Haixu Wu, Jiehui Xu, Jianmin Wang, and Mingsheng Long. 2021a. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in Neural Information Processing Systems, Vol. 34 (2021), 22419--22430.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.14778\/3503585.3503604"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.xinn.2021.100179"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.3390\/pr10040724"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1137\/1.9781611976700.60"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.energy.2022.126034"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i8.20881"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467401"},{"key":"e_1_3_2_1_44_1","volume-title":"Less is more: Fast multivariate time series forecasting with light sampling-oriented mlp structures. arXiv preprint arXiv:2207.01186","author":"Zhang Tianping","year":"2022","unstructured":"Tianping Zhang , Yizhuo Zhang , Wei Cao , Jiang Bian , Xiaohan Yi , Shun Zheng , and Jian Li. 2022b. Less is more: Fast multivariate time series forecasting with light sampling-oriented mlp structures. arXiv preprint arXiv:2207.01186 ( 2022 ). Tianping Zhang, Yizhuo Zhang, Wei Cao, Jiang Bian, Xiaohan Yi, Shun Zheng, and Jian Li. 2022b. Less is more: Fast multivariate time series forecasting with light sampling-oriented mlp structures. arXiv preprint arXiv:2207.01186 (2022)."},{"key":"e_1_3_2_1_45_1","volume-title":"First De-Trend then Attend: Rethinking Attention for Time-Series Forecasting. arXiv preprint arXiv:2212.08151","author":"Zhang Xiyuan","year":"2022","unstructured":"Xiyuan Zhang , Xiaoyong Jin , Karthick Gopalswamy , Gaurav Gupta , Youngsuk Park , Xingjian Shi , Hao Wang , Danielle C Maddix , and Yuyang Wang . 2022a. First De-Trend then Attend: Rethinking Attention for Time-Series Forecasting. arXiv preprint arXiv:2212.08151 ( 2022 ). Xiyuan Zhang, Xiaoyong Jin, Karthick Gopalswamy, Gaurav Gupta, Youngsuk Park, Xingjian Shi, Hao Wang, Danielle C Maddix, and Yuyang Wang. 2022a. First De-Trend then Attend: Rethinking Attention for Time-Series Forecasting. arXiv preprint arXiv:2212.08151 (2022)."},{"key":"e_1_3_2_1_46_1","volume-title":"The Eleventh International Conference on Learning Representations.","author":"Zhang Yunhao","year":"2023","unstructured":"Yunhao Zhang and Junchi Yan . 2023 . Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting . In The Eleventh International Conference on Learning Representations. Yunhao Zhang and Junchi Yan. 2023. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"e_1_3_2_1_48_1","first-page":"12677","article-title":"Film: Frequency improved legendre memory model for long-term time series forecasting","volume":"35","author":"Zhou Tian","year":"2022","unstructured":"Tian Zhou , Ziqing Ma , Qingsong Wen , Liang Sun , Tao Yao , Wotao Yin , and Rong Jin . 2022 a. Film: Frequency improved legendre memory model for long-term time series forecasting . Advances in Neural Information Processing Systems , Vol. 35 (2022), 12677 -- 12690 . Tian Zhou, Ziqing Ma, Qingsong Wen, Liang Sun, Tao Yao, Wotao Yin, and Rong Jin. 2022a. Film: Frequency improved legendre memory model for long-term time series forecasting. Advances in Neural Information Processing Systems, Vol. 35 (2022), 12677--12690.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_49_1","volume-title":"International Conference on Machine Learning. PMLR, 27268--27286","author":"Zhou Tian","year":"2022","unstructured":"Tian Zhou , Ziqing Ma , Qingsong Wen , Xue Wang , Liang Sun , and Rong Jin . 2022 b. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting . In International Conference on Machine Learning. PMLR, 27268--27286 . Tian Zhou, Ziqing Ma, Qingsong Wen, Xue Wang, Liang Sun, and Rong Jin. 2022b. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In International Conference on Machine Learning. PMLR, 27268--27286."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijleo.2022.170443"}],"event":{"name":"CIKM '23: The 32nd ACM International Conference on Information and Knowledge Management","location":"Birmingham United Kingdom","acronym":"CIKM '23","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web","SIGIR ACM Special Interest Group on Information Retrieval"]},"container-title":["Proceedings of the 32nd ACM International Conference on Information and Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3583780.3614851","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3583780.3614851","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:36:57Z","timestamp":1750178217000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3583780.3614851"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,21]]},"references-count":50,"alternative-id":["10.1145\/3583780.3614851","10.1145\/3583780"],"URL":"https:\/\/doi.org\/10.1145\/3583780.3614851","relation":{},"subject":[],"published":{"date-parts":[[2023,10,21]]},"assertion":[{"value":"2023-10-21","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}