{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T07:28:25Z","timestamp":1763191705220,"version":"3.45.0"},"reference-count":31,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T00:00:00Z","timestamp":1751241600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T00:00:00Z","timestamp":1751241600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,6,30]]},"DOI":"10.1109\/ijcnn64981.2025.11227793","type":"proceedings-article","created":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T18:46:15Z","timestamp":1763145975000},"page":"1-8","source":"Crossref","is-referenced-by-count":0,"title":["GLoSyformer: Global-Local Synergistic Enhanced Transformer for Multivariate Time Series Forecasting"],"prefix":"10.1109","author":[{"given":"Jiafu","family":"Zhao","sequence":"first","affiliation":[{"name":"University of Jinan,School of Information Science and Engineering,Jinan,China"}]},{"given":"Jing","family":"Zhang","sequence":"additional","affiliation":[{"name":"University of Jinan,School of Information Science and Engineering,Jinan,China"}]},{"given":"Dongmei","family":"Niu","sequence":"additional","affiliation":[{"name":"University of Jinan,School of Information Science and Engineering,Jinan,China"}]},{"given":"Tongzheng","family":"Zhu","sequence":"additional","affiliation":[{"name":"University of Jinan,School of Information Science and Engineering,Jinan,China"}]},{"given":"Chenghao","family":"Zhang","sequence":"additional","affiliation":[{"name":"University of Jinan,School of Information Science and Engineering,Jinan,China"}]},{"given":"Junzheng","family":"Yang","sequence":"additional","affiliation":[{"name":"Yantai Institute of Science and Technology,School of Data Intelligence,Yantai,China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2018.2868970"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1049\/cit2.12060"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2024.111926"},{"article-title":"ST-MLP: A cascaded spatio-temporal linear framework with channel-independence strategy for traffic forecasting","year":"2023","author":"Wang","key":"ref4"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671451"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref7","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Neural Information Processing Systems"},{"article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","year":"2020","author":"Dosovitskiy","key":"ref8"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2023\/759"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26317"},{"article-title":"iTransformer: Inverted Transformers Are Effective for Time Series Forecasting","volume-title":"The Twelfth International Conference on Learning Representations","author":"Liu","key":"ref11"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.52202\/079017-0015"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-00296-0_5"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1017\/9781108164818"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1198\/073500107000000313"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1016\/j.oceaneng.2023.115886"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2019.2935152"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-24797-2_4"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3210006"},{"article-title":"Segrnn: Segment recurrent neural network for long-term time series forecasting","year":"2023","author":"Lin","key":"ref20"},{"article-title":"Tslanet: Rethinking transformers for time series representation learning","year":"2024","author":"Eldele","key":"ref21"},{"article-title":"Micn: Multi-scale local and global context modeling for long-term series forecasting","volume-title":"The Eleventh International Conference on Learning Representations","author":"Wang","key":"ref22"},{"article-title":"TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis","volume-title":"The Eleventh International Conference on Learning Representations","author":"Wu","key":"ref23"},{"article-title":"ModernTCN: A modern pure convolution structure for general time series analysis","volume-title":"The Twelfth International Conference on Learning Representations","author":"Luo","key":"ref24"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"ref26","first-page":"22419","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","volume":"34","author":"Wu","year":"2021","journal-title":"Neural Information Processing Systems"},{"key":"ref27","first-page":"27268","article-title":"Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting","volume-title":"International Conference on Machine Learning","author":"Zhou"},{"article-title":"Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting","volume-title":"The Eleventh International Conference on Learning Representations","author":"Zhang","key":"ref28"},{"article-title":"A time series is worth 64 words: Long-term forecasting with transformers","volume-title":"The Eleventh International Conference on Learning Representations","author":"Nie","key":"ref29"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3615136"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"}],"event":{"name":"2025 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2025,6,30]]},"location":"Rome, Italy","end":{"date-parts":[[2025,7,5]]}},"container-title":["2025 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11227166\/11227148\/11227793.pdf?arnumber=11227793","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T07:25:43Z","timestamp":1763191543000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11227793\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,30]]},"references-count":31,"URL":"https:\/\/doi.org\/10.1109\/ijcnn64981.2025.11227793","relation":{},"subject":[],"published":{"date-parts":[[2025,6,30]]}}}