{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,27]],"date-time":"2025-06-27T04:06:09Z","timestamp":1750997169494,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":21,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,2,21]]},"DOI":"10.1145\/3725472.3725479","type":"proceedings-article","created":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T11:46:11Z","timestamp":1750938371000},"page":"44-50","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["MEDNet: A Mixture of Experts and Dual Sampling Framework for Multivariate Climate Time Series Prediction"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9706-3806","authenticated-orcid":false,"given":"Zeyu","family":"Zheng","sequence":"first","affiliation":[{"name":"School of Information Engineering, Huzhou University, Huzhou, Zhejiang, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-6682-6469","authenticated-orcid":false,"given":"Shujuan","family":"Zheng","sequence":"additional","affiliation":[{"name":"School of Information Engineering, Huzhou University, Huzhou, Zhejiang, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-6463-2065","authenticated-orcid":false,"given":"Xiaoshuai","family":"Zhang","sequence":"additional","affiliation":[{"name":"School of Information Engineering, Huzhou University, Huzhou, Zhejiang, China"}]}],"member":"320","published-online":{"date-parts":[[2025,6,26]]},"reference":[{"key":"e_1_3_3_1_1_2","doi-asserted-by":"crossref","unstructured":"Rahmat F R Syaputra H A Faza S et al.Prediction of Regional Revenue and Expenditure Budget using Autoregressive Integrated Moving Average (ARIMA)[J].IOP Conference Series Materials Science and Engineering 2020 851(1):012064.","DOI":"10.1088\/1757-899X\/851\/1\/012064"},{"key":"e_1_3_3_1_2_2","volume-title":"Hybrid Deep Learning Model for COVID-19 Prediction Using Convolutional Neural Network (CNN) and Bidirectional Long Short-Term Memory (LSTM) Network[J]. learning","author":"Malhotra D","year":"2023","unstructured":"Malhotra D, Sodhi G K. Hybrid Deep Learning Model for COVID-19 Prediction Using Convolutional Neural Network (CNN) and Bidirectional Long Short-Term Memory (LSTM) Network[J]. learning, 2023, 4: 7."},{"key":"e_1_3_3_1_3_2","volume-title":"Attention is All you Need[J]. Neural Information Processing Systems,Neural Information Processing Systems","year":"2017","unstructured":"VASWANI A, SHAZEER N, PARMAR N, et al. Attention is All you Need[J]. Neural Information Processing Systems,Neural Information Processing Systems, 2017."},{"key":"e_1_3_3_1_4_2","volume-title":"Do R K G, et al. Convolutional neural networks: an overview and application in radiology[J]. Insights into imaging","author":"Yamashita R","year":"2018","unstructured":"Yamashita R, Nishio M, Do R K G, et al. Convolutional neural networks: an overview and application in radiology[J]. Insights into imaging, 2018, 9: 611-629."},{"key":"e_1_3_3_1_5_2","doi-asserted-by":"crossref","unstructured":"Luzhou L Yuezhe G Bingxin C et al.Passenger Flow Scale Prediction of Urban Rail Transit Stations Based on Multilayer Perceptron (MLP)[J].Complexity 2023 2023","DOI":"10.1155\/2023\/1430449"},{"key":"e_1_3_3_1_6_2","volume-title":"AAAI Conference on Artificial Intelligence (AAAI)","author":"Haoyi Zhou","year":"2021","unstructured":"Haoyi Zhou, Shanghang Zhang, Jieqi Peng, Shuai Zhang, Jianxin Li, Hui Xiong, and Wan Zhang. Informer: Beyond efficient transformer for long sequence time-series forecasting. In AAAI Conference on Artificial Intelligence (AAAI), 2021."},{"key":"e_1_3_3_1_7_2","volume-title":"Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting[J]","year":"2021","unstructured":"WU H, XU J, WANG J, et al. Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting[J]. Cornell University- arXiv,Cornell University - arXiv, 2021."},{"key":"e_1_3_3_1_8_2","unstructured":"ZHOU T MA Z WEN Q et al. FEDformer: Frequency Enhanced Decomposed Transformer for Long-term Series Forecasting[J]."},{"key":"e_1_3_3_1_9_2","volume-title":"A Time Series is Worth 64 Words: Long-term Forecasting with Transformers[J]","author":"Nam H, SINTHONG P","year":"2022","unstructured":"NIE Y, NGUYEN NamH, SINTHONG P, et al. A Time Series is Worth 64 Words: Long-term Forecasting with Transformers[J]. 2022."},{"key":"e_1_3_3_1_10_2","volume-title":"Petformer: Long-term time series forecasting via placeholder-enhanced transformer[J]. arXiv preprint arXiv:2308.04791","author":"Lin S","year":"2023","unstructured":"Lin S, Lin W, Wu W, et al. Petformer: Long-term time series forecasting via placeholder-enhanced transformer[J]. arXiv preprint arXiv:2308.04791, 2023."},{"key":"e_1_3_3_1_11_2","first-page":"5816","article-title":"Scinet: Time series modeling and forecasting with sample convolution and interaction[J]","volume":"35","author":"Liu M","year":"2022","unstructured":"Liu M, Zeng A, Chen M, et al. Scinet: Time series modeling and forecasting with sample convolution and interaction[J]. Advances in Neural Information Processing Systems, 2022, 35: 5816-5828.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_3_1_12_2","volume-title":"TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis[J]","year":"2022","unstructured":"WU H, HU T, LIU Y, et al. TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis[J]. 2022."},{"key":"e_1_3_3_1_13_2","volume-title":"Micn: Multi-scale local and global context modeling for long-term series forecasting[C]\/\/The Eleventh International Conference on Learning Representations","author":"Wang H","year":"2022","unstructured":"Wang H, Peng J, Huang F, et al. Micn: Multi-scale local and global context modeling for long-term series forecasting[C]\/\/The Eleventh International Conference on Learning Representations. 2022."},{"key":"e_1_3_3_1_14_2","volume-title":"Are transformers effective for time series forecasting?[C]\/\/Proceedings of the AAAI conference on artificial intelligence","author":"Zeng A","year":"2023","unstructured":"Zeng A, Chen M, Zhang L, et al. Are transformers effective for time series forecasting?[C]\/\/Proceedings of the AAAI conference on artificial intelligence. 2023, 37(9): 11121-11128."},{"key":"e_1_3_3_1_15_2","volume-title":"Long-term forecasting with tide: Time-series dense encoder[J]. arXiv preprint arXiv:2304.08424","author":"Das A","year":"2023","unstructured":"Das A, Kong W, Leach A, et al. Long-term forecasting with tide: Time-series dense encoder[J]. arXiv preprint arXiv:2304.08424, 2023."},{"key":"e_1_3_3_1_16_2","volume-title":"Tsmixer: Lightweight mlp-mixer model for multivariate time series forecasting[C]\/\/Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023: 459-469","author":"Ekambaram V","unstructured":"Ekambaram V, Jati A, Nguyen N, et al. Tsmixer: Lightweight mlp-mixer model for multivariate time series forecasting[C]\/\/Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023: 459-469."},{"key":"e_1_3_3_1_17_2","volume-title":"SparseTSF: Modeling Long-term Time Series Forecasting with 1k Parameters[J]. arXiv preprint arXiv:2405.00946","author":"Lin S","year":"2024","unstructured":"Lin S, Lin W, Wu W, et al. SparseTSF: Modeling Long-term Time Series Forecasting with 1k Parameters[J]. arXiv preprint arXiv:2405.00946, 2024."},{"key":"e_1_3_3_1_18_2","volume-title":"Time-MoE: Billion-Scale Time Series Foundation Models with Mixture of Experts[J]. arXiv preprint arXiv:2409.16040","author":"Shi X","year":"2024","unstructured":"Shi X, Wang S, Nie Y, et al. Time-MoE: Billion-Scale Time Series Foundation Models with Mixture of Experts[J]. arXiv preprint arXiv:2409.16040, 2024."},{"key":"e_1_3_3_1_19_2","article-title":"Using Feed-Forward Neural Network (FFNN) In Time Series Forecasting[J]","author":"Nasser A A","year":"2009","unstructured":"Nasser A A, Rahi A K, Ahmad N M. Using Feed-Forward Neural Network (FFNN) In Time Series Forecasting[J]. Journal of Administration and Economics,(74), 2009.","journal-title":"Journal of Administration and Economics,(74)"},{"key":"e_1_3_3_1_20_2","doi-asserted-by":"crossref","unstructured":"Hu J Shen L Sun G. Squeeze-and-excitation networks[C]\/\/Proceedings of the IEEE conference on computer vision and pattern recognition. 2018: 7132-7141.","DOI":"10.1109\/CVPR.2018.00745"},{"key":"e_1_3_3_1_21_2","first-page":"1","article-title":"Root mean square error (RMSE) or mean absolute error (MAE): When to use them or not[J]","volume":"2022","author":"Hodson T O","year":"2022","unstructured":"Hodson T O. Root mean square error (RMSE) or mean absolute error (MAE): When to use them or not[J]. Geoscientific Model Development Discussions, 2022, 2022: 1-10.","journal-title":"Geoscientific Model Development Discussions"}],"event":{"name":"DSDE 2025: 2025 The 8th International Conference on Data Storage and Data Engineering","location":"Nanjing China","acronym":"DSDE 2025"},"container-title":["Proceedings of the 2025 8th International Conference on Data Storage and Data Engineering"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3725472.3725479","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T11:46:31Z","timestamp":1750938391000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3725472.3725479"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,21]]},"references-count":21,"alternative-id":["10.1145\/3725472.3725479","10.1145\/3725472"],"URL":"https:\/\/doi.org\/10.1145\/3725472.3725479","relation":{},"subject":[],"published":{"date-parts":[[2025,2,21]]},"assertion":[{"value":"2025-06-26","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}