{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,8]],"date-time":"2026-03-08T01:37:15Z","timestamp":1772933835523,"version":"3.50.1"},"reference-count":25,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,12,8]],"date-time":"2025-12-08T00:00:00Z","timestamp":1765152000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,8]],"date-time":"2025-12-08T00:00:00Z","timestamp":1765152000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,12,8]]},"DOI":"10.1109\/bigdata66926.2025.11401048","type":"proceedings-article","created":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T20:57:57Z","timestamp":1772830677000},"page":"796-801","source":"Crossref","is-referenced-by-count":0,"title":["Sig-Patchformer: A Path Signature Based Transformer for Efficient Time Series Forecasting"],"prefix":"10.1109","author":[{"given":"Rohan","family":"Akkineni","sequence":"first","affiliation":[{"name":"GITAM (Deemed to be) University,Department of Computer Science and Engineering,Hyderabad,India"}]},{"given":"Chandrasekhar","family":"Uddagiri","sequence":"additional","affiliation":[{"name":"GITAM (Deemed to be) University,Department of Artificial Intelligence and Data Science,Hyderabad,India"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-97239-3_1"},{"key":"ref2","article-title":"A generalised signature method for multivariate time series feature extraction","author":"Morrill","year":"2020","journal-title":"arXiv preprint"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"ref4","article-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting","author":"Wu","year":"2021","journal-title":"arXiv preprint"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"ref6","article-title":"A time series is worth 64 words: Long-term forecasting with transformers","author":"Nie","year":"2022","journal-title":"arXiv preprint"},{"key":"ref7","article-title":"A decoder-only foundation model for time-series forecasting","author":"Das","year":"2023","journal-title":"arXiv preprint"},{"key":"ref8","article-title":"Pathformer: Multi-scale transformers with adaptive pathways for time series forecasting","author":"Chen","year":"2024","journal-title":"arXiv preprint"},{"key":"ref9","article-title":"Timexer: Empowering transformers for time series forecasting with exogenous variables","author":"Wang","year":"2024","journal-title":"arXiv preprint"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v39i12.33378"},{"key":"ref11","article-title":"Rough transformers: Lightweight and continuous time series modelling through signature patching","author":"Moreno-Pino","year":"2024","journal-title":"arXiv preprint"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3671928"},{"key":"ref13","article-title":"Are language models actually useful for time series forecasting?","author":"Tan","year":"2024","journal-title":"arXiv preprint"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1145\/3627673.3679724"},{"key":"ref15","article-title":"Knowledgeenhanced transformer for multivariate long sequence time-series forecasting","author":"Kakde","year":"2024","journal-title":"arXiv preprint"},{"key":"ref16","article-title":"Curse of attention: A kernel-based perspective for why transformers fail to generalize on time series forecasting and beyond","author":"Ke","year":"2024","journal-title":"arXiv preprint"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2024\/608"},{"key":"ref18","first-page":"49","article-title":"Improving multi-agent reinforcement learning with stable prefix policy","volume-title":"Proceedings of the 33rd International Joint Conference on Artificial Intelligence (IJCAI-24)","author":"Deng","year":"2024"},{"key":"ref19","article-title":"Path attention: Position encoding via accumulating householder transformations","author":"Yang","year":"2025","journal-title":"arXiv preprint"},{"key":"ref20","article-title":"Powerformer: A transformer with weighted causal attention for time-series forecasting","author":"Hegazy","year":"2025","journal-title":"arXiv preprint"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"ref22","article-title":"Weather dataset: Max planck institute for biogeochemistry, germany","volume-title":"Max Planck Institute for Biogeochemistry","year":"2025"},{"key":"ref23","article-title":"Attention is all you need","author":"Vaswani","year":"2017","journal-title":"Advances in Neural Information Processing Systems (NeurIPS)"},{"key":"ref24","article-title":"Crossformer: Transformer utilizing crossdimension dependency for multivariate time series forecasting","volume-title":"International Conference on Learning Representations (ICLR)","author":"Zhang","year":"2023"},{"key":"ref25","article-title":"Reformer: The efficient transformer","author":"Kitaev","year":"2020","journal-title":"arXiv preprint"}],"event":{"name":"2025 IEEE International Conference on Big Data (BigData)","location":"Macau, China","start":{"date-parts":[[2025,12,8]]},"end":{"date-parts":[[2025,12,11]]}},"container-title":["2025 IEEE International Conference on Big Data (BigData)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11400704\/11400712\/11401048.pdf?arnumber=11401048","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T06:53:08Z","timestamp":1772866388000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11401048\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,8]]},"references-count":25,"URL":"https:\/\/doi.org\/10.1109\/bigdata66926.2025.11401048","relation":{},"subject":[],"published":{"date-parts":[[2025,12,8]]}}}