{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,19]],"date-time":"2026-02-19T07:12:21Z","timestamp":1771485141942,"version":"3.50.1"},"reference-count":20,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Sichuan Science and Technology Department Key Research and Development Project","award":["2024YFHZ0062"],"award-info":[{"award-number":["2024YFHZ0062"]}]},{"name":"Special Science Foundation of Sichuan Province for Artificial Intelligence","award":["2024ZDZX0002[16]"],"award-info":[{"award-number":["2024ZDZX0002[16]"]}]},{"name":"Special Science Foundation of Sichuan Province for Artificial Intelligence","award":["2024NSFTD0054"],"award-info":[{"award-number":["2024NSFTD0054"]}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["JBK2103012"],"award-info":[{"award-number":["JBK2103012"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Signal Process. Lett."],"published-print":{"date-parts":[[2024]]},"DOI":"10.1109\/lsp.2024.3439103","type":"journal-article","created":{"date-parts":[[2024,8,6]],"date-time":"2024-08-06T18:36:33Z","timestamp":1722969393000},"page":"2675-2679","source":"Crossref","is-referenced-by-count":7,"title":["Cross-Scale Attention for Long-Term Time Series Forecasting"],"prefix":"10.1109","volume":"31","author":[{"ORCID":"https:\/\/orcid.org\/0009-0000-3493-6403","authenticated-orcid":false,"given":"Liangjian","family":"Wen","sequence":"first","affiliation":[{"name":"School of Computing and Artificial Intelligence and the Research Institute for Digital Economy and Interdisciplinary Sciences, Southwestern University of Finance and Economics, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-2937-6131","authenticated-orcid":false,"given":"Quan","family":"Hu","sequence":"additional","affiliation":[{"name":"School of Statistics, Southwestern University of Finance and Economics, Chengdu, China"}]},{"given":"Cong","family":"Guo","sequence":"additional","affiliation":[{"name":"School of Computing and Artificial Intelligence and the Research Institute for Digital Economy and Interdisciplinary Sciences, Southwestern University of Finance and Economics, Chengdu, China"}]},{"given":"Ao","family":"Hu","sequence":"additional","affiliation":[{"name":"School of Computing and Artificial Intelligence and the Research Institute for Digital Economy and Interdisciplinary Sciences, Southwestern University of Finance and Economics, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1993-6728","authenticated-orcid":false,"given":"Mingyi","family":"Zhang","sequence":"additional","affiliation":[{"name":"Department of Biotherapy, West China Hospital, Sichuan University,, Chengdu, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2021.3058102"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2023\/233"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2022.3228131"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2023.3265330"},{"key":"ref5","article-title":"Attention is all you need","volume-title":"Proc. Int. Adv. Conf. Neural Inf. Process. Syst.","author":"Vaswani","year":"2017"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"ref7","article-title":"Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Liu","year":"2021"},{"key":"ref8","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","volume-title":"Proc. Int. Adv. Conf. Neural Inf. Process. Syst.","author":"Wu","year":"2021"},{"key":"ref9","article-title":"A time series is worth 64 words: Long-term forecasting with transformers","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Nie","year":"2023"},{"key":"ref10","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Dosovitskiy","year":"2021"},{"key":"ref11","article-title":"Language models are few-shot learners","volume-title":"Proc. Int. Adv. Conf. Neural Inf. Process. Syst.","author":"Brown","year":"2020"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"ref13","article-title":"FEDformer: Frequency enhanced decomposed transformer for long-term series forecasting","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Zhou","year":"2022"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/tkde.2024.3400008"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1609.03499"},{"key":"ref16","article-title":"Conditional time series forecasting with convolutional neural networks","author":"Borovykh","year":"2017"},{"key":"ref17","article-title":"FITS: Modeling time series with 10 k parameters","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Xu Zhijian","year":"2024"},{"key":"ref18","article-title":"Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Zhang","year":"2023"},{"key":"ref19","article-title":"itransformer: Inverted transformers are effective for time series forecasting","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Liu","year":"2024"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2022.3145908"}],"container-title":["IEEE Signal Processing Letters"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/97\/10380231\/10623694.pdf?arnumber=10623694","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,8]],"date-time":"2024-10-08T05:14:39Z","timestamp":1728364479000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10623694\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"references-count":20,"URL":"https:\/\/doi.org\/10.1109\/lsp.2024.3439103","relation":{},"ISSN":["1070-9908","1558-2361"],"issn-type":[{"value":"1070-9908","type":"print"},{"value":"1558-2361","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]}}}