{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,4]],"date-time":"2025-02-04T11:40:16Z","timestamp":1738669216627,"version":"3.36.0"},"reference-count":49,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2024,8,29]],"date-time":"2024-08-29T00:00:00Z","timestamp":1724889600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,8,29]],"date-time":"2024-08-29T00:00:00Z","timestamp":1724889600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"The National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62376081","U21A20484"],"award-info":[{"award-number":["62376081","U21A20484"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. Mach. Learn. &amp; Cyber."],"published-print":{"date-parts":[[2025,2]]},"DOI":"10.1007\/s13042-024-02317-x","type":"journal-article","created":{"date-parts":[[2024,8,29]],"date-time":"2024-08-29T10:02:41Z","timestamp":1724925761000},"page":"999-1014","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Long-term time series forecasting based on Siamese network: a perspective on few-shot learning"],"prefix":"10.1007","volume":"16","author":[{"given":"Jin","family":"Fan","sequence":"first","affiliation":[]},{"given":"Jiaqian","family":"Xiang","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Zheyu","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Huifeng","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,29]]},"reference":[{"issue":"1\u20132","key":"2317_CR1","doi-asserted-by":"publisher","first-page":"307","DOI":"10.1016\/S0925-2312(03)00372-2","volume":"55","author":"K-J Kim","year":"2003","unstructured":"Kim K-J (2003) Financial time series forecasting using support vector machines. Neurocomputing 55(1\u20132):307\u2013319","journal-title":"Neurocomputing"},{"issue":"6","key":"2317_CR2","doi-asserted-by":"publisher","first-page":"1506","DOI":"10.1109\/TNN.2003.820556","volume":"14","author":"L-J Cao","year":"2003","unstructured":"Cao L-J, Tay FEH (2003) Support vector machine with adaptive parameters in financial time series forecasting. IEEE Trans Neural Netw 14(6):1506\u20131518","journal-title":"IEEE Trans Neural Netw"},{"unstructured":"Laptev N, Yosinski J, Li LE, Smyl S (2017) Time-series extreme event forecasting with neural networks at uber. In: International conference on machine learning, vol 34, pp 1\u20135","key":"2317_CR3"},{"doi-asserted-by":"crossref","unstructured":"Zou D, Wang S, Li X, Peng H, Wang Y, Liu C, Sheng K, Zhang B (2024) Multispans: a multi-range spatial-temporal transformer network for traffic forecast via structural entropy optimization. In: Proceedings of the 17th ACM International conference on web search and data mining, pp 1032\u20131041","key":"2317_CR4","DOI":"10.1145\/3616855.3635820"},{"doi-asserted-by":"crossref","unstructured":"Yu B, Yin H, Zhu Z (2018) Spatio-temporal graph convolutional networks: a deep learning framework for traffic forecasting. In: Proceedings of the 27th International joint conference on artificial intelligence, pp 3634\u20133640","key":"2317_CR5","DOI":"10.24963\/ijcai.2018\/505"},{"unstructured":"Nie Y, Nguyen NH, Sinthong P, Kalagnanam J (2023) A time series is worth 64 words: long-term forecasting with transformers. In: International conference on learning representations (ICLR)","key":"2317_CR6"},{"doi-asserted-by":"crossref","unstructured":"Lin S, Lin W, Wu W, Wang S, Wang Y (2023) Petformer: Long-term time series forecasting via placeholder-enhanced transformer. arXiv preprint arXiv:2308.04791","key":"2317_CR7","DOI":"10.1109\/TETCI.2024.3502437"},{"doi-asserted-by":"crossref","unstructured":"Zeng A, Chen M, Zhang L, Xu Q (2023) Are transformers effective for time series forecasting? In: Proceedings of the AAAI conference on artificial intelligence, vol 37, pp 11121\u201311128","key":"2317_CR8","DOI":"10.1609\/aaai.v37i9.26317"},{"doi-asserted-by":"crossref","unstructured":"Assran M, Caron M, Misra I, Bojanowski P, Bordes F, Vincent P, Joulin A, Rabbat M, Ballas N (2022) Masked siamese networks for label-efficient learning. In: European conference on computer vision, pp 456\u2013473","key":"2317_CR9","DOI":"10.1007\/978-3-031-19821-2_26"},{"key":"2317_CR10","doi-asserted-by":"publisher","first-page":"151754","DOI":"10.1109\/ACCESS.2019.2947510","volume":"7","author":"B Wang","year":"2019","unstructured":"Wang B, Wang D (2019) Plant leaves classification: a few-shot learning method based on siamese network. Ieee Access 7:151754\u2013151763","journal-title":"Ieee Access"},{"issue":"10\u201312","key":"2317_CR11","doi-asserted-by":"publisher","first-page":"2006","DOI":"10.1016\/j.neucom.2009.09.020","volume":"73","author":"C Lemke","year":"2010","unstructured":"Lemke C, Gabrys B (2010) Meta-learning for time series forecasting and forecast combination. Neurocomputing 73(10\u201312):2006\u20132016","journal-title":"Neurocomputing"},{"issue":"18","key":"2317_CR12","first-page":"16","volume":"6","author":"TS Talagala","year":"2018","unstructured":"Talagala TS, Hyndman RJ, Athanasopoulos G et al (2018) Meta-learning how to forecast time series. Monash Econom Bus Stat Work Pap 6(18):16","journal-title":"Monash Econom Bus Stat Work Pap"},{"key":"2317_CR13","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1016\/j.procs.2018.07.204","volume":"126","author":"AR Ali","year":"2018","unstructured":"Ali AR, Gabrys B, Budka M (2018) Cross-domain meta-learning for time-series forecasting. Procedia Comput Sci 126:9\u201318","journal-title":"Procedia Comput Sci"},{"unstructured":"Xu J, Li K (2022) Automated few-shot time series forecasting based on bi-level programming. arXiv preprint arXiv:2203.03328","key":"2317_CR14"},{"unstructured":"Zhou T, Niu P, Wang X, Sun L, Jin R (2023) One fits all: Universal time series analysis by pretrained lm and specially designed adaptors. Adv Neural Inf Proc Syst","key":"2317_CR15"},{"doi-asserted-by":"crossref","unstructured":"Hooshmand A, Sharma R (2019) Energy predictive models with limited data using transfer learning. In: Proceedings of the tenth ACM International conference on future energy systems, pp 12\u201316","key":"2317_CR16","DOI":"10.1145\/3307772.3328284"},{"key":"2317_CR17","doi-asserted-by":"publisher","first-page":"352","DOI":"10.1016\/j.enbuild.2018.01.034","volume":"165","author":"M Ribeiro","year":"2018","unstructured":"Ribeiro M, Grolinger K, ElYamany HF, Higashino WA, Capretz MA (2018) Transfer learning with seasonal and trend adjustment for cross-building energy forecasting. Energy Build 165:352\u2013363","journal-title":"Energy Build"},{"key":"2317_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.energy.2021.121271","volume":"234","author":"H Yin","year":"2021","unstructured":"Yin H, Ou Z, Fu J, Cai Y, Chen S, Meng A (2021) A novel transfer learning approach for wind power prediction based on a serio-parallel deep learning architecture. Energy 234:121271","journal-title":"Energy"},{"unstructured":"Zhang Y, Yan J (2023) Crossformer: transformer utilizing cross-dimension dependency for multivariate time series forecasting. In: International conference on learning representations (ICLR)","key":"2317_CR19"},{"doi-asserted-by":"crossref","unstructured":"Zhang Y, Wu R, Dascalu SM, Harris FC (2024) Multi-scale transformer pyramid networks for multivariate time series forecasting. IEEE Access","key":"2317_CR20","DOI":"10.1109\/ACCESS.2024.3357693"},{"unstructured":"Zhang Y, Ma L, Pal S, Zhang Y, Coates M (2024) Multi-resolution time-series transformer for long-term forecasting. In: International conference on artificial intelligence and statistics, pp 4222\u20134230","key":"2317_CR21"},{"unstructured":"Gong Z, Tang Y, Liang J (2023) Patchmixer: a patch-mixing architecture for long-term time series forecasting. arXiv preprint arXiv:2310.00655","key":"2317_CR22"},{"unstructured":"Jin M, Wang S, Ma L, Chu Z, Zhang J, Shi X, Chen P-Y, Liang Y, Li Y-f, Pan S et al (2024) Time-llm: time series forecasting by reprogramming large language models. In: International conference on learning representations","key":"2317_CR23"},{"issue":"7","key":"2317_CR24","doi-asserted-by":"publisher","first-page":"1723","DOI":"10.14778\/3654621.3654637","volume":"17","author":"S Zhong","year":"2024","unstructured":"Zhong S, Song S, Zhuo W, Li G, Liu Y, Chan S-HG (2024) A multi-scale decomposition mlp-mixer for time series analysis. Proc VLDB Endow 17(7):1723\u20131736","journal-title":"Proc VLDB Endow"},{"unstructured":"Wang H, Peng J, Huang F, Wang J, Chen J, Xiao Y (2023) Micn: multi-scale local and global context modeling for long-term series forecasting. In: The eleventh international conference on learning representations","key":"2317_CR25"},{"unstructured":"Shabani MA, Abdi AH, Meng L, Sylvain T (2023) Scaleformer: iterative multi-scale refining transformers for time series forecasting. In: The eleventh international conference on learning representations","key":"2317_CR26"},{"doi-asserted-by":"crossref","unstructured":"Peng H, Zhang J, Huang X, Hao Z, Li A, Yu Z, Yu PS (2024) Unsupervised social bot detection via structural information theory. ACM Trans Inf Syst","key":"2317_CR27","DOI":"10.1145\/3660522"},{"unstructured":"Box GE, Jenkins GM, Reinsel GC, Ljung GM (2016) Time series analysis: forecasting and control. J Am Stat Assoc 68(342)","key":"2317_CR28"},{"issue":"8","key":"2317_CR29","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735\u20131780","journal-title":"Neural Comput"},{"issue":"11","key":"2317_CR30","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y LeCun","year":"1998","unstructured":"LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278\u20132324","journal-title":"Proc IEEE"},{"issue":"5\u20136","key":"2317_CR31","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1016\/0925-2312(91)90023-5","volume":"2","author":"F Murtagh","year":"1991","unstructured":"Murtagh F (1991) Multilayer perceptrons for classification and regression. Neurocomputing 2(5\u20136):183\u2013197","journal-title":"Neurocomputing"},{"unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser \u0141, Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process Syst 30","key":"2317_CR32"},{"doi-asserted-by":"crossref","unstructured":"Lai G, Chang W-C, Yang Y, Liu H (2018) Modeling long-and short-term temporal patterns with deep neural networks. In: The 41st International ACM SIGIR conference on research & development in information retrieval, pp 95\u2013104","key":"2317_CR33","DOI":"10.1145\/3209978.3210006"},{"issue":"3","key":"2317_CR34","doi-asserted-by":"publisher","first-page":"1181","DOI":"10.1016\/j.ijforecast.2019.07.001","volume":"36","author":"D Salinas","year":"2020","unstructured":"Salinas D, Flunkert V, Gasthaus J, Januschowski T (2020) Deepar: probabilistic forecasting with autoregressive recurrent networks. Int J Forecast 36(3):1181\u20131191","journal-title":"Int J Forecast"},{"unstructured":"Shi X, Chen Z, Wang H, Yeung D-Y, Wong W-K, Woo W-c (2015) Convolutional lstm network: a machine learning approach for precipitation nowcasting. Adv Neural Inf Process Syst 28","key":"2317_CR35"},{"doi-asserted-by":"crossref","unstructured":"Hossain MS, Mahmood H (2020) Short-term load forecasting using an lstm neural network. In: 2020 IEEE power and energy conference at Illinois (PECI), pp 1\u20136","key":"2317_CR36","DOI":"10.1109\/PECI48348.2020.9064654"},{"doi-asserted-by":"crossref","unstructured":"Cho K, Merrienboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using rnn encoder\u2013decoder for statistical machine translation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 1724\u20131734","key":"2317_CR37","DOI":"10.3115\/v1\/D14-1179"},{"unstructured":"Bai S, Kolter JZ, Koltun V (2018) An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. Universal language model fine-tuning for text classification","key":"2317_CR38"},{"key":"2317_CR39","first-page":"5816","volume":"35","author":"M Liu","year":"2022","unstructured":"Liu M, Zeng A, Chen M, Xu Z, Lai Q, Ma L, Xu Q (2022) Scinet: time series modeling and forecasting with sample convolution and interaction. Adv Neural Inf Process Syst 35:5816\u20135828","journal-title":"Adv Neural Inf Process Syst"},{"unstructured":"Yu F, Koltun V (2016) Multi-scale context aggregation by dilated convolutions. In: ICLR","key":"2317_CR40"},{"unstructured":"Wu H, Hu T, Liu Y, Zhou H, Wang J, Long M (2023) Timesnet: temporal 2d-variation modeling for general time series analysis. In: The eleventh international conference on learning representations","key":"2317_CR41"},{"unstructured":"Li S, Jin X, Xuan Y, Zhou X, Chen W, Wang Y-X, Yan X (2019) Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Adv Neural Inf Process Syst 32","key":"2317_CR42"},{"doi-asserted-by":"crossref","unstructured":"Zhou H, Zhang S, Peng J, Zhang S, Li J, Xiong H, Zhang W (2021) Informer: Beyond efficient transformer for long sequence time-series forecasting. In: Proceedings of the AAAI conference on artificial intelligence, vol 35, pp 11106\u201311115","key":"2317_CR43","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"2317_CR44","first-page":"22419","volume":"34","author":"H Wu","year":"2021","unstructured":"Wu H, Xu J, Wang J, Long M (2021) Autoformer: decomposition transformers with auto-correlation for long-term series forecasting. Adv Neural Inf Process Syst 34:22419\u201322430","journal-title":"Adv Neural Inf Process Syst"},{"unstructured":"Zhou T, Ma Z, Wen Q, Wang X, Sun L, Jin R (2022) Fedformer: frequency enhanced decomposed transformer for long-term series forecasting. In: International conference on machine learning, pp 27268\u201327286","key":"2317_CR45"},{"unstructured":"Woo G, Liu C, Sahoo D, Kumar A, Hoi S (2022) Etsformer: exponential smoothing transformers for time-series forecasting. arXiv preprint arXiv:2202.01381","key":"2317_CR46"},{"unstructured":"Lin S, Lin W, Wu W, Zhao F, Mo R, Zhang H (2023) Segrnn: segment recurrent neural network for long-term time series forecasting. arXiv preprint arXiv:2308.11200","key":"2317_CR47"},{"doi-asserted-by":"crossref","unstructured":"Liu X, Hu J, Li Y, Diao S, Liang Y, Hooi B, Zimmermann R (2024) Unitime: A language-empowered unified model for cross-domain time series forecasting. In: Proceedings of the ACM on web conference 2024, pp 4095\u20134106","key":"2317_CR48","DOI":"10.1145\/3589334.3645434"},{"unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S et al (2021) An image is worth 16x16 words: transformers for image recognition at scale. In: International conference on learning representations","key":"2317_CR49"}],"container-title":["International Journal of Machine Learning and Cybernetics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-024-02317-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13042-024-02317-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-024-02317-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,4]],"date-time":"2025-02-04T10:42:14Z","timestamp":1738665734000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13042-024-02317-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,29]]},"references-count":49,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["2317"],"URL":"https:\/\/doi.org\/10.1007\/s13042-024-02317-x","relation":{},"ISSN":["1868-8071","1868-808X"],"issn-type":[{"type":"print","value":"1868-8071"},{"type":"electronic","value":"1868-808X"}],"subject":[],"published":{"date-parts":[[2024,8,29]]},"assertion":[{"value":"1 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 August 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 August 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}