{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T19:19:17Z","timestamp":1742930357705,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":25,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819756650"},{"type":"electronic","value":"9789819756667"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-5666-7_22","type":"book-chapter","created":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T20:37:45Z","timestamp":1722544665000},"page":"258-268","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Generating Time Series by Using Latent Space"],"prefix":"10.1007","author":[{"given":"Xinyu","family":"Cui","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2207-0953","authenticated-orcid":false,"given":"Chunkai","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,1]]},"reference":[{"key":"22_CR1","unstructured":"ECL load. https:\/\/archive.ics.uci.edu\/ml\/datasets\/ElectricityLoadDiagrams20112014"},{"key":"22_CR2","unstructured":"Liu, S., et al.: Pyraformer: low-complexity pyramidal attention for long-range time series modeling and forecasting. In International Conference on Learning Representations, May 2021"},{"key":"22_CR3","unstructured":"Qi, X., Hou, K., Liu, T., Yu, Z., Hu, S., Ou, W.: From known to unknown: knowledge-guided transformer for time-series sales forecasting in Alibaba. arXiv preprint arXiv:2109.08381 (2021)"},{"key":"22_CR4","doi-asserted-by":"crossref","unstructured":"Zhou, H., et al.: Informer: beyond efficient transformer for long sequence time-series forecasting. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 12, pp. 11106\u201311115, May 2021","DOI":"10.1609\/aaai.v35i12.17325"},{"issue":"4","key":"22_CR5","doi-asserted-by":"publisher","first-page":"1748","DOI":"10.1016\/j.ijforecast.2021.03.012","volume":"37","author":"B Lim","year":"2021","unstructured":"Lim, B., Ar\u0131k, S.\u00d6., Loeff, N., Pfister, T.: Temporal fusion transformers for interpretable multi-horizon time series forecasting. Int. J. Forecast. 37(4), 1748\u20131764 (2021)","journal-title":"Int. J. Forecast."},{"key":"22_CR6","first-page":"22419","volume":"34","author":"H Wu","year":"2021","unstructured":"Wu, H., Xu, J., Wang, J., Long, M.: Autoformer: decomposition transformers with auto-correlation for long-term series forecasting. Adv. Neural. Inf. Process. Syst. 34, 22419\u201322430 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"3","key":"22_CR7","doi-asserted-by":"publisher","first-page":"1181","DOI":"10.1016\/j.ijforecast.2019.07.001","volume":"36","author":"D Salinas","year":"2020","unstructured":"Salinas, D., Flunkert, V., Gasthaus, J., Januschowski, T.: DeepAR: probabilistic forecasting with autoregressive recurrent networks. Int. J. Forecast. 36(3), 1181\u20131191 (2020)","journal-title":"Int. J. Forecast."},{"key":"22_CR8","first-page":"17105","volume":"33","author":"S Wu","year":"2020","unstructured":"Wu, S., Xiao, X., Ding, Q., Zhao, P., Wei, Y., Huang, J.: Adversarial sparse transformer for time series forecasting. Adv. Neural. Inf. Process. Syst. 33, 17105\u201317115 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"22_CR9","unstructured":"Li, S., et al.: Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Adv. Neural Inf. Process. Syst. 32 (2019)"},{"key":"22_CR10","unstructured":"Zhou, T., Ma, Z., Wen, Q., Wang, X., Sun, L., Jin, R.: Fedformer: frequency enhanced decomposed transformer for long-term series forecasting. In: International Conference on Machine Learning, pp. 27268\u201327286. PMLR, June 2022"},{"key":"22_CR11","doi-asserted-by":"crossref","unstructured":"Lin, Y., Koprinska, I., Rana, M.: SSDNet: state space decomposition neural network for time series forecasting. In: 2021 IEEE International Conference on Data Mining (ICDM), pp. 370\u2013378. IEEE, December 2021","DOI":"10.1109\/ICDM51629.2021.00048"},{"key":"22_CR12","first-page":"23592","volume":"34","author":"B Tang","year":"2021","unstructured":"Tang, B., Matteson, D.S.: Probabilistic transformer for time series analysis. Adv. Neural. Inf. Process. Syst. 34, 23592\u201323608 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"2","key":"22_CR13","doi-asserted-by":"publisher","first-page":"240","DOI":"10.1109\/72.279188","volume":"5","author":"JT Connor","year":"1994","unstructured":"Connor, J.T., Martin, R.D., Atlas, L.E.: Recurrent neural networks and robust time series prediction. IEEE Trans. Neural Netw. 5(2), 240\u2013254 (1994)","journal-title":"IEEE Trans. Neural Netw."},{"issue":"4","key":"22_CR14","doi-asserted-by":"publisher","first-page":"8013","DOI":"10.3233\/JIFS-189623","volume":"40","author":"HM Choi","year":"2021","unstructured":"Choi, H.M., Kim, M.K., Yang, H.: Abnormally high water temperature prediction using LSTM deep learning model. J. Intell. Fuzzy Syst. 40(4), 8013\u20138020 (2021)","journal-title":"J. Intell. Fuzzy Syst."},{"key":"22_CR15","doi-asserted-by":"publisher","first-page":"821","DOI":"10.1007\/s12145-020-00462-9","volume":"13","author":"J Becerra-Rico","year":"2020","unstructured":"Becerra-Rico, J., Aceves-Fern\u00e1ndez, M.A., Esquivel-Escalante, K., Pedraza-Ortega, J.C.: Airborne particle pollution predictive model using Gated Recurrent Unit (GRU) deep neural networks. Earth Sci. Inf. 13, 821\u2013834 (2020)","journal-title":"Earth Sci. Inf."},{"key":"22_CR16","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2021\/5589075","volume":"2021","author":"RL Abduljabbar","year":"2021","unstructured":"Abduljabbar, R.L., Dia, H., Tsai, P.W.: Unidirectional and bidirectional LSTM models for short-term traffic prediction. J. Adv. Transp. 2021, 1\u201316 (2021)","journal-title":"J. Adv. Transp."},{"key":"22_CR17","unstructured":"Dama, F., Sinoquet, C.: Time series analysis and modeling to forecast: a survey. arXiv preprint arXiv:2104.00164 (2021)"},{"key":"22_CR18","unstructured":"Kitaev, N., Kaiser, L., Levskaya, A.: Reformer: the efficient transformer. In: International Conference on Learning Representations, September 2019"},{"key":"22_CR19","doi-asserted-by":"crossref","unstructured":"Lai, G., Chang, W.C., Yang, Y., Liu, H.: Modeling long-and short-term temporal patterns with deep neural networks. In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 95\u2013104, June 2018","DOI":"10.1145\/3209978.3210006"},{"key":"22_CR20","doi-asserted-by":"crossref","unstructured":"Cheng, Z., Yang, Y., Wang, W., Hu, W., Zhuang, Y., Song, G.: Time2graph: revisiting time series modeling with dynamic shapelets. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 04, 3617\u20133624, April 2020","DOI":"10.1609\/aaai.v34i04.5769"},{"key":"22_CR21","first-page":"9881","volume":"35","author":"Y Liu","year":"2022","unstructured":"Liu, Y., Wu, H., Wang, J., Long, M.: Non-stationary transformers: exploring the stationarity in time series forecasting. Adv. Neural. Inf. Process. Syst. 35, 9881\u20139893 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"22_CR22","doi-asserted-by":"crossref","unstructured":"Lara-Ben\u00edtez, P., Carranza-Garc\u00eda, M., Luna-Romera, J.M., Riquelme, J.C.: Temporal convolutional networks applied to energy-related time series forecasting. Appl. Sci. 10(7), 2322 (2020)","DOI":"10.3390\/app10072322"},{"key":"22_CR23","doi-asserted-by":"crossref","unstructured":"Hu, W., Yang, Y., Cheng, Z., Yang, C., Ren, X.: Time-series event prediction with evolutionary state graph. In: Proceedings of the 14th ACM International Conference on Web Search and Data Mining, pp. 580\u2013588","DOI":"10.1145\/3437963.3441827"},{"key":"22_CR24","doi-asserted-by":"crossref","unstructured":"Nguyen, N., Quanz, B.: Temporal latent auto-encoder: a method for probabilistic multivariate time series forecasting. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 10, pp. 9117\u20139125, May 2021","DOI":"10.1609\/aaai.v35i10.17101"},{"key":"22_CR25","doi-asserted-by":"publisher","first-page":"302","DOI":"10.1016\/j.neucom.2018.12.084","volume":"396","author":"Z Shen","year":"2020","unstructured":"Shen, Z., Zhang, Y., Lu, J., Xu, J., Xiao, G.: A novel time series forecasting model with deep learning. Neurocomputing 396, 302\u2013313 (2020)","journal-title":"Neurocomputing"}],"container-title":["Lecture Notes in Computer Science","Advanced Intelligent Computing Technology and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-5666-7_22","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,25]],"date-time":"2024-11-25T16:24:44Z","timestamp":1732551884000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-5666-7_22"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819756650","9789819756667"],"references-count":25,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-5666-7_22","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"1 August 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"This work was supported by Natural Science Foundation of Guangdong Province, China (Grant NO. 2024A1515010242), Guangdong Provincial Key Laboratory of Novel Security Intelligence Technologies, China (Grant NO. 2022B1212010005) and Shenzhen Research Council (Grant NO. WD20220811170253002).","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Funding Statement"}},{"value":"ICIC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tianjin","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 August 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 August 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icic2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.ic-icc.cn\/2024\/index.htm","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}