{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T05:07:50Z","timestamp":1735708070861,"version":"3.32.0"},"reference-count":76,"publisher":"Association for Computing Machinery (ACM)","issue":"12","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. VLDB Endow."],"published-print":{"date-parts":[[2024,8]]},"abstract":"<jats:p>Forecasting extrapolates the values of a time series into the future, and is crucial to optimize core operations for many businesses and organizations. Building machine learning (ML)-based forecasting applications presents a challenge though, due to non-stationary data and large numbers of time series. As there is no single dominating approach to forecasting, forecasting systems have to support a wide variety of approaches, ranging from deep learning-based methods to classical methods built on probabilistic modelling.<\/jats:p>\n          <jats:p>\n            We revisit our earlier work on a monolithic platform for forecasting from VLDB 2017, and describe how we evolved it into a modern forecasting stack consisting of several layers that support a wide range of forecasting needs and automate common tasks like model selection. This stack leverages our open source forecasting libraries\n            <jats:italic>GluonTS<\/jats:italic>\n            and\n            <jats:italic>AutoGluon-TimeSeries<\/jats:italic>\n            , the scalable ML platform\n            <jats:italic>SageMaker<\/jats:italic>\n            , and forms the basis of the no-code forecasting solutions (\n            <jats:italic>SageMaker Canvas<\/jats:italic>\n            and\n            <jats:italic>Amazon Forecast<\/jats:italic>\n            ), available in the Amazon Web Services cloud. We give insights into the predictive performance of our stack and discuss learnings from using it to provision resources for the cloud database services DynamoDB, Redshift and Athena.\n          <\/jats:p>","DOI":"10.14778\/3685800.3685813","type":"journal-article","created":{"date-parts":[[2024,11,8]],"date-time":"2024-11-08T17:25:21Z","timestamp":1731086721000},"page":"3883-3892","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["A Flexible Forecasting Stack"],"prefix":"10.14778","volume":"17","author":[{"given":"Tim","family":"Januschowski","sequence":"first","affiliation":[{"name":"Databricks"}]},{"given":"Yuyang","family":"Wang","sequence":"additional","affiliation":[{"name":"Amazon"}]},{"given":"Jan","family":"Gasthaus","sequence":"additional","affiliation":[]},{"given":"Syama","family":"Rangapuram","sequence":"additional","affiliation":[{"name":"Amazon"}]},{"given":"Caner","family":"T\u00fcrkmen","sequence":"additional","affiliation":[{"name":"Amazon"}]},{"given":"Jasper","family":"Zschiegner","sequence":"additional","affiliation":[{"name":"Amazon"}]},{"given":"Lorenzo","family":"Stella","sequence":"additional","affiliation":[{"name":"Amazon"}]},{"given":"Michael","family":"Bohlke-Schneider","sequence":"additional","affiliation":[{"name":"Amazon"}]},{"given":"Danielle","family":"Maddix","sequence":"additional","affiliation":[{"name":"Amazon"}]},{"given":"Konstantinos","family":"Benidis","sequence":"additional","affiliation":[{"name":"Amazon"}]},{"given":"Alexander","family":"Alexandrov","sequence":"additional","affiliation":[{"name":"Materialize"}]},{"given":"Christos","family":"Faloutsos","sequence":"additional","affiliation":[{"name":"Amazon &amp; CMU"}]},{"given":"Sebastian","family":"Schelter","sequence":"additional","affiliation":[{"name":"BIFOLD &amp; TU Berlin"}]}],"member":"320","published-online":{"date-parts":[[2024,11,8]]},"reference":[{"key":"e_1_2_1_1_1","volume-title":"Ali Caner T\u00fcrkmen, and Yuyang Wang","author":"Alexandrov Alexander","year":"2019","unstructured":"Alexander Alexandrov, Konstantinos Benidis, Michael Bohlke-Schneider, Valentin Flunkert, Jan Gasthaus, Tim Januschowski, Danielle C Maddix, Syama Rangapuram, David Salinas, Jasper Schulz, Lorenzo Stella, Ali Caner T\u00fcrkmen, and Yuyang Wang. 2019. GluonTS: Probabilistic Time Series Models in Python. arXiv preprint arXiv:1906.05264 (2019)."},{"key":"e_1_2_1_2_1","volume-title":"Retrieved","author":"Services Amazon Web","year":"2023","unstructured":"Amazon Web Services. 2023. CNN-QR Algorithm. Retrieved July 16, 2024 from https:\/\/docs.aws.amazon.com\/forecast\/latest\/dg\/aws-forecast-algo-cnnqr.html"},{"key":"e_1_2_1_3_1","volume-title":"Sebastian Pineda Arango, Shubham Kapoor, et al.","author":"Ansari Abdul Fatir","year":"2024","unstructured":"Abdul Fatir Ansari, Lorenzo Stella, Caner Turkmen, Xiyuan Zhang, Pedro Mercado, Huibin Shen, Oleksandr Shchur, Syama Sundar Rangapuram, Sebastian Pineda Arango, Shubham Kapoor, et al. 2024. Chronos: Learning the language of time series. arXiv preprint arXiv:2403.07815 (2024)."},{"key":"e_1_2_1_4_1","volume-title":"Zakaria Haque, Salem Haykal, Mustafa Ispir, Vihan Jain, Levent Koc, et al.","author":"Baylor Denis","year":"2017","unstructured":"Denis Baylor, Eric Breck, Heng-Tze Cheng, Noah Fiedel, Chuan Yu Foo, Zakaria Haque, Salem Haykal, Mustafa Ispir, Vihan Jain, Levent Koc, et al. 2017. Tfx: A tensorflow-based production-scale machine learning platform. KDD (2017), 1387--1395."},{"key":"e_1_2_1_5_1","volume-title":"Retrieved","author":"Beitner Jan","year":"2024","unstructured":"Jan Beitner. 2024. PyTorch Forecasting. Retrieved July 16, 2024 from https:\/\/pytorch-forecasting.readthedocs.io\/en\/stable\/"},{"key":"e_1_2_1_6_1","volume-title":"Syama Sundar Rangapuram","author":"Benidis Konstantinos","year":"2020","unstructured":"Konstantinos Benidis, Syama Sundar Rangapuram, Valentin Flunkert, Bernie Wang, Danielle Maddix, Caner Turkmen, Jan Gasthaus, Michael Bohlke-Schneider, David Salinas, Lorenzo Stella, Laurent Callot, and Tim Januschowski. 2020. Neural forecasting: Introduction and literature overview. arXiv:2004.10240"},{"key":"e_1_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3533382"},{"key":"e_1_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.14778\/3137765.3137775"},{"key":"e_1_2_1_9_1","volume-title":"Computation reuse via fusion in Amazon Athena. ICDE","author":"Bruno Nicolas","year":"2022","unstructured":"Nicolas Bruno, Johnny Debrodt, Chujun Song, and Wei Zheng. 2022. Computation reuse via fusion in Amazon Athena. ICDE (2022), 1610--1620."},{"key":"e_1_2_1_10_1","volume-title":"Ensemble selection from libraries of models. ICML","author":"Caruana Rich","year":"2004","unstructured":"Rich Caruana, Alexandru Niculescu-Mizil, Geoff Crew, and Alex Ksikes. 2004. Ensemble selection from libraries of models. ICML (2004), 18."},{"key":"e_1_2_1_11_1","volume-title":"NeurIPS Workshop on Machine Learning Systems","author":"Chen Tianqi","year":"2015","unstructured":"Tianqi Chen, Mu Li, Yutian Li, Min Lin, Naiyan Wang, Minjie Wang, Tianjun Xiao, Bing Xu, Chiyuan Zhang, and Zheng Zhang. 2015. Mxnet: A flexible and efficient machine learning library for heterogeneous distributed systems. NeurIPS Workshop on Machine Learning Systems (2015)."},{"key":"e_1_2_1_12_1","volume-title":"A decoder-only foundation model for time-series forecasting. arXiv:2310.10688","author":"Das Abhimanyu","year":"2023","unstructured":"Abhimanyu Das, Weihao Kong, Rajat Sen, and Yichen Zhou. 2023. A decoder-only foundation model for time-series forecasting. arXiv:2310.10688 (2023)."},{"key":"e_1_2_1_13_1","volume-title":"Chirag Mohapatra, Siddartha Naidu, and Colin White.","author":"Dooley Samuel","year":"2023","unstructured":"Samuel Dooley, Gurnoor Singh Khurana, Chirag Mohapatra, Siddartha Naidu, and Colin White. 2023. ForecastPFN: Synthetically-Trained Zero-Shot Forecasting. NeurIPS (2023)."},{"key":"e_1_2_1_14_1","volume-title":"MQTransformer: Multi-Horizon Forecasts with Context Dependent and Feedback-Aware Attention. arXiv preprint arXiv:2009.14799","author":"Eisenach Carson","year":"2020","unstructured":"Carson Eisenach, Yagna Patel, and Dhruv Madeka. 2020. MQTransformer: Multi-Horizon Forecasts with Context Dependent and Feedback-Aware Attention. arXiv preprint arXiv:2009.14799 (2020)."},{"key":"e_1_2_1_15_1","volume-title":"Autogluon-tabular: Robust and accurate automl for structured data. arXiv preprint arXiv:2003.06505","author":"Erickson Nick","year":"2020","unstructured":"Nick Erickson, Jonas Mueller, Alexander Shirkov, Hang Zhang, Pedro Larroy, Mu Li, and Alexander Smola. 2020. Autogluon-tabular: Robust and accurate automl for structured data. arXiv preprint arXiv:2003.06505 (2020)."},{"key":"e_1_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.14778\/3229863.3229878"},{"key":"e_1_2_1_17_1","unstructured":"Azul Garza and Max Mergenthaler-Canseco. 2023. TimeGPT-1. arXiv:2310.03589"},{"key":"e_1_2_1_18_1","volume-title":"Retrieved","author":"Garza Federico Garza","year":"2024","unstructured":"Federico Garza Garza, Max Mergenthaler Canseco, Cristian Chall\u00fa, and Kin G. Olivares. 2022. StatsForecast: Lightning fast forecasting with statistical and econometric models. Retrieved July 16, 2024 from https:\/\/github.com\/Nixtla\/statsforecast"},{"key":"e_1_2_1_19_1","volume-title":"AISTATS (2019)","author":"Gasthaus Jan","year":"2019","unstructured":"Jan Gasthaus, Konstantinos Benidis, Yuyang Wang, Syama Sundar Rangapuram, David Salinas, Valentin Flunkert, and Tim Januschowski. 2019. Probabilistic Forecasting with Spline Quantile Function RNNs. AISTATS (2019), 1901--1910."},{"key":"e_1_2_1_20_1","volume-title":"The dynamic factor analysis of economic time series. Latent variables in socio-economic models","author":"Geweke John","year":"1977","unstructured":"John Geweke. 1977. The dynamic factor analysis of economic time series. Latent variables in socio-economic models (1977)."},{"key":"e_1_2_1_21_1","first-page":"1","article-title":"Amlb: an automl benchmark","volume":"25","author":"Gijsbers Pieter","year":"2024","unstructured":"Pieter Gijsbers, Marcos LP Bueno, Stefan Coors, Erin LeDell, S\u00e9bastien Poirier, Janek Thomas, Bernd Bischl, and Joaquin Vanschoren. 2024. Amlb: an automl benchmark. Journal of Machine Learning Research 25, 101 (2024), 1--65.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1467-9868.2007.00587.x"},{"key":"e_1_2_1_23_1","unstructured":"Rakshitha Godahewa Christoph Bergmeir Geoffrey I. Webb Rob J. Hyndman and Pablo Montero-Manso. 2021. Monash Time Series Forecasting Archive. arXiv:2105.06643"},{"key":"e_1_2_1_24_1","volume-title":"MOMENT: A Family of Open Time-series Foundation Models. arXiv preprint arXiv:2402.03885","author":"Goswami Mononito","year":"2024","unstructured":"Mononito Goswami, Konrad Szafer, Arjun Choudhry, Yifu Cai, Shuo Li, and Artur Dubrawski. 2024. MOMENT: A Family of Open Time-series Foundation Models. arXiv preprint arXiv:2402.03885 (2024)."},{"key":"e_1_2_1_25_1","unstructured":"Nate Gruver Marc Finzi Shikai Qiu and Andrew Gordon Wilson. 2023. Large Language Models Are Zero-Shot Time Series Forecasters. In Advances in Neural Information Processing Systems."},{"key":"e_1_2_1_26_1","volume-title":"Large language models are zero-shot time series forecasters. NeurIPS 36","author":"Gruver Nate","year":"2024","unstructured":"Nate Gruver, Marc Finzi, Shikai Qiu, and Andrew G Wilson. 2024. Large language models are zero-shot time series forecasters. NeurIPS 36 (2024)."},{"key":"e_1_2_1_27_1","first-page":"6404","article-title":"Probabilistic forecasting: A level-set approach","volume":"34","author":"Hasson Hilaf","year":"2021","unstructured":"Hilaf Hasson, Bernie Wang, Tim Januschowski, and Jan Gasthaus. 2021. Probabilistic forecasting: A level-set approach. NeurIPS 34 (2021), 6404--6416.","journal-title":"NeurIPS"},{"key":"e_1_2_1_28_1","first-page":"1","article-title":"Darts: User-Friendly Modern Machine Learning for Time Series","volume":"23","author":"Herzen Julien","year":"2022","unstructured":"Julien Herzen, Francesco Lassig, Samuele Giuliano Piazzetta, Thomas Neuer, Lao Tafti, Guillaume Raille, Tomas Van Pottelbergh, Marek Pasieka, Andrzej Skrodzki, Nicolas Huguenin, Maxime Dumonal, Jan Koacisz, Dennis Bader, Frederick Gusset, Mounir Benheddi, Camila Williamson, Michal Kosinski, Matej Petrik, and Gael Grosch. 2022. Darts: User-Friendly Modern Machine Learning for Time Series. Journal of Machine Learning Research 23, 124 (2022), 1--6.","journal-title":"Journal of Machine Learning Research"},{"volume-title":"Forecasting with exponential smoothing: the state space approach","author":"Hyndman Rob","key":"e_1_2_1_29_1","unstructured":"Rob Hyndman, Anne B Koehler, J Keith Ord, and Ralph D Snyder. 2008. Forecasting with exponential smoothing: the state space approach. Springer Science & Business Media."},{"key":"e_1_2_1_30_1","unstructured":"Rob J Hyndman and George Athanasopoulos. 2018. Forecasting: principles and practice. OTexts."},{"key":"e_1_2_1_31_1","volume-title":"Automatic time series forecasting: the forecast package for R. Journal of Statistical Software","author":"Hyndman Rob J","year":"2008","unstructured":"Rob J Hyndman and Yeasmin Khandakar. 2008. Automatic time series forecasting: the forecast package for R. Journal of Statistical Software (2008)."},{"key":"e_1_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2006.03.001"},{"key":"e_1_2_1_33_1","first-page":"20","article-title":"Open-Source Forecasting Tools in Python. Foresight","volume":"55","author":"Januschowski Tim","year":"2019","unstructured":"Tim Januschowski, Jan Gasthaus, and Yuyang Wang. 2019. Open-Source Forecasting Tools in Python. Foresight: The International Journal of Applied Forecasting 55 (2019), 20--26.","journal-title":"The International Journal of Applied Forecasting"},{"key":"e_1_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2019.05.008"},{"key":"e_1_2_1_35_1","first-page":"36","article-title":"A Classification of Business Forecasting Problems. Foresight","volume":"52","author":"Januschowski Tim","year":"2019","unstructured":"Tim Januschowski and Stephan Kolassa. 2019. A Classification of Business Forecasting Problems. Foresight: The International Journal of Applied Forecasting 52 (2019), 36--43.","journal-title":"The International Journal of Applied Forecasting"},{"key":"e_1_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2021.10.004"},{"key":"e_1_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2021.07.004"},{"key":"e_1_2_1_38_1","volume-title":"Time-LLM: Time Series Forecasting by Reprogramming Large Language Models. ICLR","author":"Jin Ming","year":"2024","unstructured":"Ming Jin, Shiyu Wang, Lintao Ma, Zhixuan Chu, James Y. Zhang, Xiaoming Shi, Pin-Yu Chen, Yuxuan Liang, Yuan-Fang Li, Shirui Pan, and Qingsong Wen. 2024. Time-LLM: Time Series Forecasting by Reprogramming Large Language Models. ICLR (2024)."},{"key":"e_1_2_1_39_1","doi-asserted-by":"crossref","unstructured":"Manuel Kunz Stefan Birr Mones Raslan Lei Ma Zhen Li Adele Gouttes Mateusz Koren Tofigh Naghibi Johannes Stephan Mariia Bulycheva Matthias Grzeschik Armin Keki\u0107 Michael Narodovitch Kashif Rasul Julian Sieber and Tim Januschowski. 2023. Deep Learning based Forecasting: a case study from the online fashion industry. arXiv:2305.14406","DOI":"10.1007\/978-3-031-35879-1_11"},{"key":"e_1_2_1_40_1","first-page":"15371","article-title":"Deep Rao-Blackwellised Particle Filters for Time Series Forecasting","volume":"33","author":"Kurle Richard","year":"2020","unstructured":"Richard Kurle, Syama Sundar Rangapuram, Emmanuel de B\u00e9zenac, Stephan G\u00fcnnemann, and Jan Gasthaus. 2020. Deep Rao-Blackwellised Particle Filters for Time Series Forecasting. NeurIPS 33 (2020), 15371--15382.","journal-title":"NeurIPS"},{"key":"e_1_2_1_41_1","volume-title":"Elastic Machine Learning Algorithms in Amazon SageMaker. SIGMOD","author":"Liberty Edo","year":"2020","unstructured":"Edo Liberty, Zohar Karnin, Bing Xiang, Laurence Rouesnel, Baris Coskun, Ramesh Nallapati, Julio Delgado, Amir Sadoughi, Amir Astashonok, Piali Das, Can Balioglu, Saswata Charkravarty, Madhav Jha, Philip Gaultier, Tim Januschowski, Valentin Flunkert, Bernie Wang, Jan Gasthaus, Syama Rangapuram, David Salinas, Sebastian Schelter, David Arpin, and Alexander Smola. 2020. Elastic Machine Learning Algorithms in Amazon SageMaker. SIGMOD (2020), 731--737."},{"key":"e_1_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2021.03.012"},{"key":"e_1_2_1_43_1","volume-title":"Retrieved","author":"Ma Zhixing","year":"2020","unstructured":"Zhixing Ma, Srinivas Chakravarthi Thandu, Rohit Menon, Vivek Ramamoorthy, and Bernie Wang. 2020. Automating your Amazon Forecast workflow with Lambda, Step Functions, and CloudWatch Events rule. Retrieved July 16, 2024 from https:\/\/aws.amazon.com\/blogs\/machine-learning\/automating-your-amazon-forecast-workflow-with-lambda-step-functions-and-cloudwatch-events-rule\/"},{"key":"e_1_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1016\/S0169-2070(00)00057-1"},{"key":"e_1_2_1_45_1","first-page":"33","article-title":"The M5 Competition and the Future of Human Expertise in Forecasting. Foresight","volume":"60","author":"Makridakis Spyros","year":"2021","unstructured":"Spyros Makridakis and Evangelos Spiliotis. 2021. The M5 Competition and the Future of Human Expertise in Forecasting. Foresight: The International Journal of Applied Forecasting 60 (2021), 33--37.","journal-title":"The International Journal of Applied Forecasting"},{"key":"e_1_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2018.06.001"},{"key":"e_1_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2019.04.014"},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2021.11.013"},{"key":"e_1_2_1_49_1","volume-title":"NBEATS: Neural basis expansion analysis for interpretable time series forecasting. ICLR","author":"Oreshkin Boris N","year":"2020","unstructured":"Boris N Oreshkin, Dmitri Carpov, Nicolas Chapados, and Yoshua Bengio. 2020. NBEATS: Neural basis expansion analysis for interpretable time series forecasting. ICLR (2020)."},{"key":"e_1_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.14778\/3476311.3476391"},{"key":"e_1_2_1_51_1","volume-title":"PyTorch: An Imperative Style","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. NeurIPS (2019), 8024--8035."},{"key":"e_1_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.5555\/1953048.2078195"},{"key":"e_1_2_1_53_1","volume-title":"Barbara Pogorzelska, Miroslav Miladinovic, Krishnaram Kenthapadi, Matthias Seeger, and C\u00e9dric Archambeau.","author":"Perrone Valerio","year":"2021","unstructured":"Valerio Perrone, Huibin Shen, Aida Zolic, Iaroslav Shcherbatyi, Amr Ahmed, Tanya Bansal, Michele Donini, Fela Winkelmolen, Rodolphe Jenatton, Jean Baptiste Faddoul, Barbara Pogorzelska, Miroslav Miladinovic, Krishnaram Kenthapadi, Matthias Seeger, and C\u00e9dric Archambeau. 2021. Amazon SageMaker Automatic Model Tuning: Scalable Gradient-Free Optimization. arXiv:2012.08489"},{"key":"e_1_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2021.11.001"},{"key":"e_1_2_1_55_1","volume-title":"Deep State Space Models for Time Series Forecasting. NeurIPS 31","author":"Rangapuram Syama Sundar","year":"2018","unstructured":"Syama Sundar Rangapuram, Matthias W Seeger, Jan Gasthaus, Lorenzo Stella, Yuyang Wang, and Tim Januschowski. 2018. Deep State Space Models for Time Series Forecasting. NeurIPS 31 (2018)."},{"key":"e_1_2_1_56_1","unstructured":"Syama Sundar Rangapuram Matthias W Seeger Jan Gasthaus Lorenzo Stella Yuyang Wang and Tim Januschowski. 2018. Deep state space models for time series forecasting. In Advances in Neural Information Processing Systems. 7785--7794."},{"key":"e_1_2_1_57_1","volume-title":"Konstantinos Benidis, Jan Gasthaus, and Tim Januschowski.","author":"Rangapuram Syama Sundar","year":"2021","unstructured":"Syama Sundar Rangapuram, Lucien Werner, Pedro Mercado Lopez, Konstantinos Benidis, Jan Gasthaus, and Tim Januschowski. 2021. End-to-End Learning of Coherent Probabilistic Forecasts for Hierarchical Time Series. ICML (2021), 8832--8843."},{"key":"e_1_2_1_58_1","unstructured":"Kashif Rasul Arjun Ashok Andrew Robert Williams Arian Khorasani George Adamopoulos Rishika Bhagwatkar Marin Bilo\u0161 Hena Ghonia Nadhir Vincent Hassen Anderson Schneider Sahil Garg Alexandre Drouin Nicolas Chapados Yuriy Nevmyvaka and Irina Rish. 2023. Lag-Llama: Towards Foundation Models for Time Series Forecasting. arXiv:2310.08278"},{"key":"e_1_2_1_59_1","volume-title":"Autoregressive Denoising Diffusion Models for Multivariate Probabilistic Time Series Forecasting. ICML","author":"Rasul Kashif","year":"2021","unstructured":"Kashif Rasul, Calvin Seward, Ingmar Schuster, and Roland Vollgraf. 2021. Autoregressive Denoising Diffusion Models for Multivariate Probabilistic Time Series Forecasting. ICML (2021), 8857--8868."},{"key":"e_1_2_1_60_1","volume-title":"Multi-variate probabilistic time series forecasting via conditioned normalizing flows. ICLR","author":"Rasul Kashif","year":"2021","unstructured":"Kashif Rasul, Abdul-Saboor Sheikh, Ingmar Schuster, Urs Bergmann, and Roland Vollgraf. 2021. Multi-variate probabilistic time series forecasting via conditioned normalizing flows. ICLR (2021)."},{"key":"e_1_2_1_61_1","volume-title":"High-dimensional multivariate forecasting with low-rank Gaussian Copula Processes. NeurIPS 32","author":"Salinas David","year":"2019","unstructured":"David Salinas, Michael Bohlke-Schneider, Laurent Callot, Roberto Medico, and Jan Gasthaus. 2019. High-dimensional multivariate forecasting with low-rank Gaussian Copula Processes. NeurIPS 32 (2019)."},{"key":"e_1_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2019.07.001"},{"key":"e_1_2_1_63_1","volume-title":"On challenges in machine learning model management","author":"Schelter Sebastian","year":"2018","unstructured":"Sebastian Schelter, Felix Biessmann, Tim Januschowski, David Salinas, Stephan Seufert, and Gyuri Szarvas. 2018. On challenges in machine learning model management. IEEE Data Engineering Bulletin (2018)."},{"key":"e_1_2_1_64_1","volume-title":"Bayesian intermittent demand forecasting for large inventories. NeurIPS","author":"Seeger Matthias W","year":"2016","unstructured":"Matthias W Seeger, David Salinas, and Valentin Flunkert. 2016. Bayesian intermittent demand forecasting for large inventories. NeurIPS (2016), 4646--4654."},{"key":"e_1_2_1_65_1","volume-title":"AutoGluon-TimeSeries: AutoML for Probabilistic Time Series Forecasting. arXiv preprint arXiv:2308.05566","author":"Shchur Oleksandr","year":"2023","unstructured":"Oleksandr Shchur, Caner Turkmen, Nick Erickson, Huibin Shen, Alexander Shirkov, Tony Hu, and Yuyang Wang. 2023. AutoGluon-TimeSeries: AutoML for Probabilistic Time Series Forecasting. arXiv preprint arXiv:2308.05566 (2023)."},{"key":"e_1_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1145\/2213836.2213945"},{"key":"e_1_2_1_67_1","volume-title":"Defuse: Harnessing Unrestricted Adversarial Examples for Debugging Models Beyond Test Accuracy. arXiv:2102.06162","author":"Slack Dylan","year":"2021","unstructured":"Dylan Slack, Nathalie Rauschmayr, and Krishnaram Kenthapadi. 2021. Defuse: Harnessing Unrestricted Adversarial Examples for Debugging Models Beyond Test Accuracy. arXiv:2102.06162"},{"key":"e_1_2_1_68_1","volume-title":"Retrieved","author":"Smyl S","year":"2018","unstructured":"S Smyl, J Ranganathan, and A Pasqua. 2018. M4 Forecasting Competition: Introducing a New Hybrid ES-RNN Model. Retrieved July 16, 2024 from https:\/\/eng.uber.com\/m4-forecasting-competition"},{"key":"e_1_2_1_69_1","volume-title":"Probabilistic Gradient Boosting Machines for Large-Scale Probabilistic Regression. KDD","author":"Sprangers Olivier","year":"2021","unstructured":"Olivier Sprangers, Sebastian Schelter, and Maarten de Rijke. 2021. Probabilistic Gradient Boosting Machines for Large-Scale Probabilistic Regression. KDD (2021), 1510--1520."},{"key":"e_1_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijforecast.2021.11.011"},{"key":"e_1_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1080\/00031305.2017.1380080"},{"key":"e_1_2_1_72_1","volume-title":"Forecasting intermittent and sparse time series: A unified probabilistic framework via deep renewal processes. PlosOne","author":"Turkmen Ali Caner","year":"2019","unstructured":"Ali Caner Turkmen, Yuyang Wang, and Tim Januschowski. 2019. Forecasting intermittent and sparse time series: A unified probabilistic framework via deep renewal processes. PlosOne (2019)."},{"key":"e_1_2_1_73_1","volume-title":"Deep factors for forecasting. ICML","author":"Wang Yuyang","year":"2019","unstructured":"Yuyang Wang, Alex Smola, Danielle Maddix, Jan Gasthaus, Dean Foster, and Tim Januschowski. 2019. Deep factors for forecasting. ICML (2019), 6607--6617."},{"key":"e_1_2_1_74_1","volume-title":"A multi-horizon quantile recurrent forecaster. arXiv preprint arXiv:1711.11053","author":"Wen Ruofeng","year":"2017","unstructured":"Ruofeng Wen, Kari Torkkola, Balakrishnan Narayanaswamy, and Dhruv Madeka. 2017. A multi-horizon quantile recurrent forecaster. arXiv preprint arXiv:1711.11053 (2017)."},{"key":"e_1_2_1_75_1","volume-title":"Unified training of universal time series forecasting transformers. arXiv preprint arXiv:2402.02592","author":"Woo Gerald","year":"2024","unstructured":"Gerald Woo, Chenghao Liu, Akshat Kumar, Caiming Xiong, Silvio Savarese, and Doyen Sahoo. 2024. Unified training of universal time series forecasting transformers. arXiv preprint arXiv:2402.02592 (2024)."},{"key":"e_1_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1145\/2934664"}],"container-title":["Proceedings of the VLDB Endowment"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.14778\/3685800.3685813","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,31]],"date-time":"2024-12-31T05:34:43Z","timestamp":1735623283000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.14778\/3685800.3685813"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8]]},"references-count":76,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["10.14778\/3685800.3685813"],"URL":"https:\/\/doi.org\/10.14778\/3685800.3685813","relation":{},"ISSN":["2150-8097"],"issn-type":[{"type":"print","value":"2150-8097"}],"subject":[],"published":{"date-parts":[[2024,8]]},"assertion":[{"value":"2024-11-08","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}