{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T01:06:09Z","timestamp":1770339969125,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":60,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,7,20]],"date-time":"2025-07-20T00:00:00Z","timestamp":1752969600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100003009","name":"Science and Technology Development Fund","doi-asserted-by":"publisher","award":["0123\/2023\/RIA2, 001\/2024\/SKL"],"award-info":[{"award-number":["0123\/2023\/RIA2, 001\/2024\/SKL"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100003009","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,7,20]]},"DOI":"10.1145\/3690624.3709260","type":"proceedings-article","created":{"date-parts":[[2025,4,4]],"date-time":"2025-04-04T18:44:43Z","timestamp":1743792283000},"page":"295-306","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":4,"title":["IN-Flow: Instance Normalization Flow for Non-stationary Time Series Forecasting"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7656-445X","authenticated-orcid":false,"given":"Wei","family":"Fan","sequence":"first","affiliation":[{"name":"University of Oxford, Oxford, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-7355-7090","authenticated-orcid":false,"given":"Shun","family":"Zheng","sequence":"additional","affiliation":[{"name":"Microsoft Research Asia, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3961-5523","authenticated-orcid":false,"given":"Pengyang","family":"Wang","sequence":"additional","affiliation":[{"name":"University of Macau, Macao SAR, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9769-8131","authenticated-orcid":false,"given":"Rui","family":"Xie","sequence":"additional","affiliation":[{"name":"University of Central Florida, Orlando, Florida, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9980-6033","authenticated-orcid":false,"given":"Kun","family":"Yi","sequence":"additional","affiliation":[{"name":"State Information Center of China, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1037-1361","authenticated-orcid":false,"given":"Qi","family":"Zhang","sequence":"additional","affiliation":[{"name":"Tongji University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9472-600X","authenticated-orcid":false,"given":"Jiang","family":"Bian","sequence":"additional","affiliation":[{"name":"Microsoft Research Asia, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1767-8024","authenticated-orcid":false,"given":"Yanjie","family":"Fu","sequence":"additional","affiliation":[{"name":"Arizona State University, Tempe, Arizona, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,7,20]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.protcy.2012.05.047"},{"key":"e_1_3_2_2_2_1","volume-title":"Grey prediction with rolling mechanism for electricity demand forecasting of Turkey. energy","author":"Akay Diyar","year":"2007","unstructured":"Diyar Akay and Mehmet Atak. 2007. Grey prediction with rolling mechanism for electricity demand forecasting of Turkey. energy, Vol. 32, 9 (2007), 1670--1675."},{"key":"e_1_3_2_2_3_1","volume-title":"Analyzing inverse problems with invertible neural networks. arXiv preprint arXiv:1808.04730","author":"Ardizzone Lynton","year":"2018","unstructured":"Lynton Ardizzone, Jakob Kruse, Sebastian Wirkert, Daniel Rahner, Eric W Pellegrini, Ralf S Klessen, Lena Maier-Hein, Carsten Rother, and Ullrich K''othe. 2018. Analyzing inverse problems with invertible neural networks. arXiv preprint arXiv:1808.04730 (2018)."},{"key":"e_1_3_2_2_4_1","volume-title":"An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling. CoRR","author":"Bai Shaojie","year":"2018","unstructured":"Shaojie Bai, J. Zico Kolter, and Vladlen Koltun. 2018. An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling. CoRR, Vol. abs\/1803.01271 (2018)."},{"key":"e_1_3_2_2_5_1","volume-title":"N-HiTS: Neural Hierarchical Interpolation for Time Series Forecasting. CoRR","author":"Challu Cristian","year":"2022","unstructured":"Cristian Challu, Kin G. Olivares, Boris N. Oreshkin, Federico Garza, Max Mergenthaler, and Artur Dubrawski. 2022. N-HiTS: Neural Hierarchical Interpolation for Time Series Forecasting. CoRR, Vol. abs\/2201.12886 (2022)."},{"key":"e_1_3_2_2_6_1","volume-title":"Gradient-based bi-level optimization for deep learning: A survey. arXiv preprint arXiv:2207.11719","author":"Chen Can","year":"2022","unstructured":"Can Chen, Xi Chen, Chen Ma, Zixuan Liu, and Xue Liu. 2022. Gradient-based bi-level optimization for deep learning: A survey. arXiv preprint arXiv:2207.11719 (2022)."},{"key":"e_1_3_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10479-007-0176-2"},{"key":"e_1_3_2_2_8_1","volume-title":"Disentangling Structured Components: Towards Adaptive, Interpretable and Scalable Time Series Forecasting","author":"Deng Jinliang","year":"2024","unstructured":"Jinliang Deng, Xiusi Chen, Renhe Jiang, Du Yin, Yi Yang, Xuan Song, and Ivor W Tsang. 2024a. Disentangling Structured Components: Towards Adaptive, Interpretable and Scalable Time Series Forecasting. IEEE Transactions on Knowledge and Data Engineering (2024)."},{"key":"e_1_3_2_2_9_1","volume-title":"Decomposition Delivers Both in Long-term Time Series Forecasting. In The Thirty-eighth Annual Conference on Neural Information Processing Systems. https:\/\/openreview.net\/forum?id=wiEHZSV15I","author":"Deng Jinliang","year":"2024","unstructured":"Jinliang Deng, Feiyang Ye, Du Yin, Xuan Song, Ivor Tsang, and Hui Xiong. 2024b. Parsimony or Capability? Decomposition Delivers Both in Long-term Time Series Forecasting. In The Thirty-eighth Annual Conference on Neural Information Processing Systems. https:\/\/openreview.net\/forum?id=wiEHZSV15I"},{"key":"e_1_3_2_2_10_1","volume-title":"Nice: Non-linear independent components estimation. arXiv preprint arXiv:1410.8516","author":"Dinh Laurent","year":"2014","unstructured":"Laurent Dinh, David Krueger, and Yoshua Bengio. 2014. Nice: Non-linear independent components estimation. arXiv preprint arXiv:1410.8516 (2014)."},{"key":"e_1_3_2_2_11_1","volume-title":"Density estimation using real nvp. arXiv preprint arXiv:1605.08803","author":"Dinh Laurent","year":"2016","unstructured":"Laurent Dinh, Jascha Sohl-Dickstein, and Samy Bengio. 2016. Density estimation using real nvp. arXiv preprint arXiv:1605.08803 (2016)."},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3482315"},{"key":"e_1_3_2_2_13_1","volume-title":"A learned representation for artistic style. arXiv preprint arXiv:1610.07629","author":"Dumoulin Vincent","year":"2016","unstructured":"Vincent Dumoulin, Jonathon Shlens, and Manjunath Kudlur. 2016. A learned representation for artistic style. arXiv preprint arXiv:1610.07629 (2016)."},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i6.25914"},{"key":"e_1_3_2_2_15_1","volume-title":"DEPTS: Deep Expansion Learning for Periodic Time Series Forecasting. In International Conference on Learning Representations.","author":"Fan Wei","year":"2022","unstructured":"Wei Fan, Shun Zheng, Xiaohan Yi, Wei Cao, Yanjie Fu, Jiang Bian, and Tie-Yan Liu. 2022. DEPTS: Deep Expansion Learning for Periodic Time Series Forecasting. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_16_1","volume-title":"International Conference on Machine Learning. PMLR, 1568--1577","author":"Franceschi Luca","year":"2018","unstructured":"Luca Franceschi, Paolo Frasconi, Saverio Salzo, Riccardo Grazzi, and Massimiliano Pontil. 2018. Bilevel programming for hyperparameter optimization and meta-learning. In International Conference on Machine Learning. PMLR, 1568--1577."},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i5.16529"},{"key":"e_1_3_2_2_18_1","volume-title":"ONR Memorandum","volume":"52","author":"Holt Charles C","year":"1957","unstructured":"Charles C Holt. 1957. Forecasting trends and seasonal by exponentially weighted moving averages. ONR Memorandum, Vol. 52, 2 (1957)."},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2003.09.015"},{"key":"e_1_3_2_2_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.167"},{"key":"e_1_3_2_2_21_1","volume-title":"International conference on machine learning. PMLR, 448--456","author":"Ioffe Sergey","year":"2015","unstructured":"Sergey Ioffe and Christian Szegedy. 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International conference on machine learning. PMLR, 448--456."},{"key":"e_1_3_2_2_22_1","volume-title":"International Conference on Learning Representations.","author":"Kim Taesung","year":"2022","unstructured":"Taesung Kim, Jinhee Kim, Yunwon Tae, Cheonbok Park, Jang-Ho Choi, and Jaegul Choo. 2022. Reversible instance normalization for accurate time-series forecasting against distribution shift. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_23_1","volume-title":"Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980","author":"Kingma Diederik P","year":"2014","unstructured":"Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)."},{"key":"e_1_3_2_2_24_1","volume-title":"Glow: Generative flow with invertible 1x1 convolutions. Advances in neural information processing systems","author":"Kingma Durk P","year":"2018","unstructured":"Durk P Kingma and Prafulla Dhariwal. 2018. Glow: Generative flow with invertible 1x1 convolutions. Advances in neural information processing systems, Vol. 31 (2018)."},{"key":"e_1_3_2_2_25_1","volume-title":"Simon JD Prince, and Marcus A Brubaker","author":"Kobyzev Ivan","year":"2020","unstructured":"Ivan Kobyzev, Simon JD Prince, and Marcus A Brubaker. 2020. Normalizing flows: An introduction and review of current methods. IEEE transactions on pattern analysis and machine intelligence, Vol. 43, 11 (2020), 3964--3979."},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"crossref","unstructured":"Guokun Lai Wei-Cheng Chang Yiming Yang and Hanxiao Liu. 2018. Modeling Long-and Short-Term Temporal Patterns with Deep Neural Networks. In SIGIR. 95--104.","DOI":"10.1145\/3209978.3210006"},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1098\/rsta.2020.0209"},{"key":"e_1_3_2_2_28_1","volume-title":"DARTS: Differentiable Architecture Search. In International Conference on Learning Representations.","author":"Liu Hanxiao","year":"2018","unstructured":"Hanxiao Liu, Karen Simonyan, and Yiming Yang. 2018. DARTS: Differentiable Architecture Search. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_29_1","first-page":"5816","article-title":"SCINet: time series modeling and forecasting with sample convolution and interaction","volume":"35","author":"Liu Minhao","year":"2022","unstructured":"Minhao Liu, Ailing Zeng, Muxi Chen, Zhijian Xu, Qiuxia Lai, Lingna Ma, and Qiang Xu. 2022b. SCINet: time series modeling and forecasting with sample convolution and interaction. Advances in Neural Information Processing Systems, Vol. 35 (2022), 5816--5828.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_30_1","volume-title":"iTransformer: Inverted Transformers Are Effective for Time Series Forecasting. CoRR","author":"Liu Yong","year":"2023","unstructured":"Yong Liu, Tengge Hu, Haoran Zhang, Haixu Wu, Shiyu Wang, Lintao Ma, and Mingsheng Long. 2023b. iTransformer: Inverted Transformers Are Effective for Time Series Forecasting. CoRR, Vol. abs\/2310.06625 (2023)."},{"key":"e_1_3_2_2_31_1","volume-title":"Non-stationary Transformers: Exploring the Stationarity in Time Series Forecasting. In Advances in Neural Information Processing Systems.","author":"Liu Yong","year":"2022","unstructured":"Yong Liu, Haixu Wu, Jianmin Wang, and Mingsheng Long. 2022a. Non-stationary Transformers: Exploring the Stationarity in Time Series Forecasting. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_2_32_1","volume-title":"Adaptive Normalization for Non-stationary Time Series Forecasting: A Temporal Slice Perspective. In Thirty-seventh Conference on Neural Information Processing Systems.","author":"Liu Zhiding","year":"2023","unstructured":"Zhiding Liu, Mingyue Cheng, Zhi Li, Zhenya Huang, Qi Liu, Yanhu Xie, and Enhong Chen. 2023a. Adaptive Normalization for Non-stationary Time Series Forecasting: A Temporal Slice Perspective. In Thirty-seventh Conference on Neural Information Processing Systems."},{"key":"e_1_3_2_2_33_1","first-page":"865","article-title":"Traffic flow prediction with big data: a deep learning approach","volume":"16","author":"Lv Yisheng","year":"2014","unstructured":"Yisheng Lv, Yanjie Duan, Wenwen Kang, Zhengxi Li, and Fei-Yue Wang. 2014. Traffic flow prediction with big data: a deep learning approach. IEEE Transactions on Intelligent Transportation Systems, Vol. 16, 2 (2014), 865--873.","journal-title":"IEEE Transactions on Intelligent Transportation Systems"},{"key":"e_1_3_2_2_34_1","unstructured":"LIU Minhao Ailing Zeng Muxi Chen Zhijian Xu LAI Qiuxia Lingna Ma and Qiang Xu. 2022. SCINet: Time Series Modeling and Forecasting with Sample Convolution and Interaction. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_2_35_1","volume-title":"International Conference on Learning Representations.","author":"Nie Yuqi","year":"2023","unstructured":"Yuqi Nie, Nam H. Nguyen, Phanwadee Sinthong, and Jayant Kalagnanam. 2023. A Time Series is Worth 64 Words: Long-term Forecasting with Transformers. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2010.5596746"},{"key":"e_1_3_2_2_37_1","volume-title":"International Conference on Learning Representations.","author":"Oreshkin Boris N.","year":"2020","unstructured":"Boris N. Oreshkin, Dmitri Carpov, Nicolas Chapados, and Yoshua Bengio. 2020. N-BEATS: Neural basis expansion analysis for interpretable time series forecasting. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_38_1","first-page":"1","article-title":"Normalizing Flows for Probabilistic Modeling and Inference","volume":"22","author":"Papamakarios George","year":"2021","unstructured":"George Papamakarios, Eric T Nalisnick, Danilo Jimenez Rezende, Shakir Mohamed, and Balaji Lakshminarayanan. 2021. Normalizing Flows for Probabilistic Modeling and Inference. J. Mach. Learn. Res., Vol. 22, 57 (2021), 1--64.","journal-title":"J. Mach. Learn. Res."},{"key":"e_1_3_2_2_39_1","volume-title":"Masked autoregressive flow for density estimation. Advances in neural information processing systems","author":"Papamakarios George","year":"2017","unstructured":"George Papamakarios, Theo Pavlakou, and Iain Murray. 2017. Masked autoregressive flow for density estimation. Advances in neural information processing systems, Vol. 30 (2017)."},{"key":"e_1_3_2_2_40_1","volume-title":"Deep adaptive input normalization for time series forecasting","author":"Passalis Nikolaos","year":"2019","unstructured":"Nikolaos Passalis, Anastasios Tefas, Juho Kanniainen, Moncef Gabbouj, and Alexandros Iosifidis. 2019. Deep adaptive input normalization for time series forecasting. IEEE transactions on neural networks and learning systems, Vol. 31, 9 (2019), 3760--3765."},{"key":"e_1_3_2_2_41_1","volume-title":"Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. 2019. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, Vol. 32 (2019)."},{"key":"e_1_3_2_2_42_1","volume-title":"Meta-learning with implicit gradients. Advances in neural information processing systems","author":"Rajeswaran Aravind","year":"2019","unstructured":"Aravind Rajeswaran, Chelsea Finn, Sham M Kakade, and Sergey Levine. 2019. Meta-learning with implicit gradients. Advances in neural information processing systems, Vol. 32 (2019)."},{"key":"e_1_3_2_2_43_1","volume-title":"Deep state space models for time series forecasting. Advances in neural information processing systems","author":"Rangapuram Syama Sundar","year":"2018","unstructured":"Syama Sundar Rangapuram, Matthias W Seeger, Jan Gasthaus, Lorenzo Stella, Yuyang Wang, and Tim Januschowski. 2018. Deep state space models for time series forecasting. Advances in neural information processing systems, Vol. 31 (2018), 7785--7794."},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2019.07.001"},{"key":"e_1_3_2_2_45_1","volume-title":"International Conference on Machine Learning. PMLR, 8741--8751","author":"Shen Sheng","year":"2020","unstructured":"Sheng Shen, Zhewei Yao, Amir Gholami, Michael Mahoney, and Kurt Keutzer. 2020. Powernorm: Rethinking batch normalization in transformers. In International Conference on Machine Learning. PMLR, 8741--8751."},{"key":"e_1_3_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1109\/TEVC.2017.2712906"},{"key":"e_1_3_2_2_47_1","volume-title":"Instance normalization: The missing ingredient for fast stylization. arXiv preprint arXiv:1607.08022","author":"Ulyanov Dmitry","year":"2016","unstructured":"Dmitry Ulyanov, Andrea Vedaldi, and Victor Lempitsky. 2016. Instance normalization: The missing ingredient for fast stylization. arXiv preprint arXiv:1607.08022 (2016)."},{"key":"e_1_3_2_2_48_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information processing systems. 5998--6008."},{"key":"e_1_3_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2014.08.008"},{"key":"e_1_3_2_2_50_1","unstructured":"Peter Whittle. 1951. Hypothesis testing in time series analysis. Almqvist & Wiksells boktr."},{"key":"e_1_3_2_2_51_1","volume-title":"Prediction and regulation by linear least-square methods","author":"Whittle Peter","unstructured":"Peter Whittle. 1963. Prediction and regulation by linear least-square methods. English Universities Press."},{"key":"e_1_3_2_2_52_1","volume-title":"Forecasting sales by exponentially weighted moving averages. Management science","author":"Winters Peter R","year":"1960","unstructured":"Peter R Winters. 1960. Forecasting sales by exponentially weighted moving averages. Management science, Vol. 6, 3 (1960), 324--342."},{"key":"e_1_3_2_2_53_1","volume-title":"Deeptime: Deep time-index meta-learning for non-stationary time-series forecasting. arXiv preprint arXiv:2207.06046","author":"Woo Gerald","year":"2022","unstructured":"Gerald Woo, Chenghao Liu, Doyen Sahoo, Akshat Kumar, and Steven Hoi. 2022. Deeptime: Deep time-index meta-learning for non-stationary time-series forecasting. arXiv preprint arXiv:2207.06046 (2022)."},{"key":"e_1_3_2_2_54_1","first-page":"22419","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","volume":"34","author":"Wu Haixu","year":"2021","unstructured":"Haixu Wu, Jiehui Xu, Jianmin Wang, and Mingsheng Long. 2021. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in Neural Information Processing Systems, Vol. 34 (2021), 22419--22430.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_55_1","volume-title":"FilterNet: Harnessing Frequency Filters for Time Series Forecasting. In The Thirty-eighth Annual Conference on Neural Information Processing Systems. https:\/\/openreview.net\/forum?id=ugL2D9idAD","author":"Yi Kun","year":"2024","unstructured":"Kun Yi, Jingru Fei, Qi Zhang, Hui He, Shufeng Hao, Defu Lian, and Wei Fan. 2024. FilterNet: Harnessing Frequency Filters for Time Series Forecasting. In The Thirty-eighth Annual Conference on Neural Information Processing Systems. https:\/\/openreview.net\/forum?id=ugL2D9idAD"},{"key":"e_1_3_2_2_56_1","unstructured":"Kun Yi Qi Zhang Wei Fan Shoujin Wang Pengyang Wang Hui He Ning An Defu Lian Longbing Cao and Zhendong Niu. 2023. Frequency-domain MLPs are More Effective Learners in Time Series Forecasting. In NeurIPS."},{"key":"e_1_3_2_2_57_1","volume-title":"Are Transformers Effective for Time Series Forecasting? arXiv preprint arXiv:2205.13504","author":"Zeng Ailing","year":"2022","unstructured":"Ailing Zeng, Muxi Chen, Lei Zhang, and Qiang Xu. 2022. Are Transformers Effective for Time Series Forecasting? arXiv preprint arXiv:2205.13504 (2022)."},{"key":"e_1_3_2_2_58_1","volume-title":"Are Transformers Effective for Time Series Forecasting?","author":"Zeng Ailing","unstructured":"Ailing Zeng, Muxi Chen, Lei Zhang, and Qiang Xu. 2023. Are Transformers Effective for Time Series Forecasting?. In AAAI. AAAI Press, 11121--11128."},{"key":"e_1_3_2_2_59_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"e_1_3_2_2_60_1","volume-title":"FEDformer: Frequency enhanced decomposed transformer for long-term series forecasting. arXiv preprint arXiv:2201.12740","author":"Zhou Tian","year":"2022","unstructured":"Tian Zhou, Ziqing Ma, Qingsong Wen, Xue Wang, Liang Sun, and Rong Jin. 2022. FEDformer: Frequency enhanced decomposed transformer for long-term series forecasting. arXiv preprint arXiv:2201.12740 (2022)."}],"event":{"name":"KDD '25: The 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Toronto ON Canada","acronym":"KDD '25","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V.1"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3690624.3709260","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3690624.3709260","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,16]],"date-time":"2025-08-16T15:43:34Z","timestamp":1755359014000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3690624.3709260"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,20]]},"references-count":60,"alternative-id":["10.1145\/3690624.3709260","10.1145\/3690624"],"URL":"https:\/\/doi.org\/10.1145\/3690624.3709260","relation":{},"subject":[],"published":{"date-parts":[[2025,7,20]]},"assertion":[{"value":"2025-07-20","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}