{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T06:10:01Z","timestamp":1751523001998,"version":"3.41.0"},"reference-count":21,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,5,12]],"date-time":"2025-05-12T00:00:00Z","timestamp":1747008000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,5,12]],"date-time":"2025-05-12T00:00:00Z","timestamp":1747008000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,5,12]]},"DOI":"10.1109\/iwcmc65282.2025.11059684","type":"proceedings-article","created":{"date-parts":[[2025,7,2]],"date-time":"2025-07-02T17:42:13Z","timestamp":1751478133000},"page":"49-54","source":"Crossref","is-referenced-by-count":0,"title":["An Efficient Timer-Expanded Multi-variate Time Series Forecasting Foundation Model"],"prefix":"10.1109","author":[{"given":"Chang","family":"Lu","sequence":"first","affiliation":[{"name":"Beijing University of Posts and Telecommunications,Beijing,China"}]},{"given":"Yuan","family":"Chang","sequence":"additional","affiliation":[{"name":"China Telecom Research Institute,Beijing,China"}]},{"given":"Aobo","family":"Liang","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications,Beijing,China"}]},{"given":"Hong","family":"Luo","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications,Beijing,China"}]},{"given":"Yan","family":"Sun","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications,Beijing,China"}]}],"member":"263","reference":[{"doi-asserted-by":"publisher","key":"ref1","DOI":"10.1016\/b978-0-444-62731-5.00016-6"},{"doi-asserted-by":"publisher","key":"ref2","DOI":"10.1038\/s41597-020-0548-x"},{"doi-asserted-by":"publisher","key":"ref3","DOI":"10.3906\/elk-1101-1029"},{"key":"ref4","article-title":"Attention is all you need[J]","author":"Vaswani","year":"2017","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref5","first-page":"32","article-title":"Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting[J]","author":"Li","year":"2019","journal-title":"Advances in neural information processing systems"},{"year":"2020","author":"Kitaev","article-title":"Reformer: The efficient transformer[J]","key":"ref6"},{"year":"2022","author":"Liu","journal-title":"Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting[C]","key":"ref7"},{"key":"ref8","first-page":"22419","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting[J]","volume":"34","author":"Wu","year":"2021","journal-title":"Advances in neural information processing systems"},{"key":"ref9","first-page":"27268","article-title":"Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting[C]","volume-title":"International conference on machine learning","author":"Zhou"},{"volume-title":"Forty-first International Conference on Machine Learning","author":"Das","article-title":"A decoder-only foundation model for timeseries forecasting[C]","key":"ref10"},{"volume-title":"Forty-first International Conference on Machine Learning","author":"Liu","article-title":"Timer: Generative Pre-trained Transformers Are Large Time Series Models[C]","key":"ref11"},{"year":"2024","author":"Woo","article-title":"Unified training of universal time series forecasting transformers[J]","key":"ref12"},{"doi-asserted-by":"publisher","key":"ref13","DOI":"10.1609\/aaai.v35i12.17325"},{"year":"2022","author":"Nie","article-title":"A time series is worth 64 words: Long-term forecasting with transformers[J]","key":"ref14"},{"volume-title":"The eleventh international conference on learning representations","author":"Zhang","article-title":"Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting[C]","key":"ref15"},{"year":"2024","author":"Wang","article-title":"Timexer: Empowering transformers for time series forecasting with exogenous variables[J]","key":"ref16"},{"year":"2024","author":"Chen","article-title":"Pathformer: Multi-scale transformers with adaptive pathways for time series forecasting[J]","key":"ref17"},{"doi-asserted-by":"publisher","key":"ref18","DOI":"10.1109\/JIOT.2024.3363451"},{"doi-asserted-by":"publisher","key":"ref19","DOI":"10.1016\/j.neunet.2024.106196"},{"year":"2024","author":"Goswami","article-title":"Moment: A family of open time-series foundation models[J]","key":"ref20"},{"year":"2024","author":"Ansari","article-title":"Chronos: Learning the language of time series[J]","key":"ref21"}],"event":{"name":"2025 International Wireless Communications and Mobile Computing (IWCMC)","start":{"date-parts":[[2025,5,12]]},"location":"Abu Dhabi, United Arab Emirates","end":{"date-parts":[[2025,5,16]]}},"container-title":["2025 International Wireless Communications and Mobile Computing (IWCMC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11059430\/11059431\/11059684.pdf?arnumber=11059684","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T05:35:29Z","timestamp":1751520929000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11059684\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,12]]},"references-count":21,"URL":"https:\/\/doi.org\/10.1109\/iwcmc65282.2025.11059684","relation":{},"subject":[],"published":{"date-parts":[[2025,5,12]]}}}