{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T20:50:07Z","timestamp":1775249407912,"version":"3.50.1"},"reference-count":33,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T00:00:00Z","timestamp":1747958400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T00:00:00Z","timestamp":1747958400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,5,23]]},"DOI":"10.1109\/cniot65435.2025.11070962","type":"proceedings-article","created":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T17:41:09Z","timestamp":1752514869000},"page":"1-7","source":"Crossref","is-referenced-by-count":3,"title":["LTSMiTransformer: Learnable Temporal Sparsity and Memory for Efficient Long-Term Time Series Forecasting"],"prefix":"10.1109","author":[{"given":"Polycarp Shizawaliyi","family":"Yakoi","sequence":"first","affiliation":[{"name":"Liaoning Technical University,School of Electronic and Information Engineering,Huludao,China"}]},{"given":"Xiangfu","family":"Meng","sequence":"additional","affiliation":[{"name":"Liaoning Technical University,School of Electronic and Information Engineering,Huludao,China"}]},{"given":"Chunli","family":"Yu","sequence":"additional","affiliation":[{"name":"Liaoning Technical University,School of Electronic and Information Engineering,Huludao,China"}]},{"given":"Victor Adeyi","family":"Odeh","sequence":"additional","affiliation":[{"name":"University of Electronic Science and Technology of China,School of Information and Communication Engineering,Chengdu,China"}]},{"given":"Yongqin","family":"Zhang","sequence":"additional","affiliation":[{"name":"Liaoning Technical University,School of Electronic and Information Engineering,Huludao,China"}]},{"given":"Zheng","family":"Zhao","sequence":"additional","affiliation":[{"name":"Liaoning Technical University,School of Electronic and Information Engineering,Huludao,China"}]},{"given":"Danladi","family":"Suleman","sequence":"additional","affiliation":[{"name":"University of the Sunshine Coast UniSC Moreton Bay, QLD,School of Science, Technology and Engineering,Australia"}]},{"given":"Yuzhuo","family":"Yang","sequence":"additional","affiliation":[{"name":"Liaoning Technical University,School of Electronic and Information Engineering,Huludao,China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"crossref","first-page":"429","DOI":"10.1016\/B978-0-12-385208-3.00014-6","article-title":"Chapter 14 - Time Series: Understanding Changes over Time","volume-title":"Practical Business Statistics","author":"Siegel","year":"2012"},{"key":"ref2","author":"Bommasani","year":"2021","journal-title":"On the opportunities and risks of foundation models"},{"key":"ref3","article-title":"Attention is all you need","volume-title":"31st Conference on Neural Information Processing Systems (NIPS 2017)","author":"Vaswani"},{"key":"ref4","volume-title":"What Is a Transformer Model?","author":"Merritt","year":"2022"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"ref6","doi-asserted-by":"crossref","first-page":"395","DOI":"10.1016\/B978-0-12-814761-0.00012-5","article-title":"Chapter 12 - Time Series Forecasting","volume-title":"Data Science","author":"Kotu","year":"2019"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1179"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-89929-5_5"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-49409-8_7"},{"key":"ref11","article-title":"Time series transformer","volume-title":"Hugging Face","author":"Rasul"},{"key":"ref12","first-page":"32","article-title":"Enhancing the Locality and Breaking the Memory Bottleneck of Transformer on Time Series Forecasting","author":"Li","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref13","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","volume-title":"35th Conference on Neural Information Processing Systems (NeurIPS 2021)","author":"J"},{"key":"ref14","author":"Nie","year":"2022","journal-title":"A time series is worth 64 words: Long-term forecasting with transformers"},{"key":"ref15","author":"Dosovitskiy","year":"2020","journal-title":"An image is worth 16x16 words: Transformers for image recognition at scale"},{"key":"ref16","author":"Kitaev","year":"2020","journal-title":"Reformer: The efficient transformer"},{"key":"ref17","author":"Child","year":"2019","journal-title":"Generating long sequences with sparse transformers"},{"key":"ref18","author":"Beltagy","year":"2020","journal-title":"Longformer: The long-document transformer"},{"key":"ref19","article-title":"Big bird: Transformers for longer sequences","author":"Zaheer","year":"2020","journal-title":"Advances in Neural Information Processing Systems 33 (NeurIPS 2020)"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/0-387-32520-4_15"},{"key":"ref21","article-title":"Meta-Learning with Memory-Augmented Neural Networks","volume-title":"33rd International Conference on Machine Learning","author":"Santoro"},{"key":"ref22","author":"Burtsev","year":"2020","journal-title":"Memory transformer"},{"key":"ref23","article-title":"itransformer: Inverted transformers are effective for time series forecasting","volume-title":"ICLR 2024","author":"Liu"},{"key":"ref24","article-title":"Cross-entropy loss functions: Theoretical analysis and applications","volume-title":"International conference on Machine learning","author":"Mao"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-32520-6_22"},{"key":"ref26","author":"Dereich","year":"2024","journal-title":"Convergence rates for the Adam optimizer"},{"key":"ref27","article-title":"Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In International conference on machine learning","volume-title":"International conference on machine learning","author":"Zhou"},{"key":"ref28","first-page":"9881","article-title":"Non-stationary transformers: Exploring the stationarity in time series forecasting","volume":"35","author":"Liu","year":"2022","journal-title":"Advances in neural information processing systems"},{"key":"ref29","article-title":"Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting","volume-title":"The eleventh international conference on learning representations","author":"Zhang"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"ref31","author":"Das","year":"2023","journal-title":"Long-term forecasting with tide: Time series dense encoder. arXiv 2023"},{"key":"ref32","first-page":"5816","article-title":"Scinet: Time series modeling and forecasting with sample convolution and interaction","volume":"35","author":"Liu","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref33","author":"Wu","year":"2022","journal-title":"Timesnet: Temporal 2d-variation modeling for general time series analysis"}],"event":{"name":"2025 6th International Conference on Computing, Networks and Internet of Things (CNIOT)","location":"Shanghai, China","start":{"date-parts":[[2025,5,23]]},"end":{"date-parts":[[2025,5,25]]}},"container-title":["2025 6th International Conference on Computing, Networks and Internet of Things (CNIOT)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11069441\/11069432\/11070962.pdf?arnumber=11070962","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T19:49:41Z","timestamp":1775245781000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11070962\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,23]]},"references-count":33,"URL":"https:\/\/doi.org\/10.1109\/cniot65435.2025.11070962","relation":{},"subject":[],"published":{"date-parts":[[2025,5,23]]}}}