{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T00:46:29Z","timestamp":1765500389611,"version":"3.48.0"},"publisher-location":"New York, NY, USA","reference-count":40,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,11,10]]},"DOI":"10.1145\/3746252.3760803","type":"proceedings-article","created":{"date-parts":[[2025,11,10]],"date-time":"2025-11-10T18:37:32Z","timestamp":1762799852000},"page":"4675-4679","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Integrating Time Series into LLMs via Multi-layer Steerable Embedding Fusion for Enhanced Forecasting"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-4406-7947","authenticated-orcid":false,"given":"Zhuomin","family":"Chen","sequence":"first","affiliation":[{"name":"School of Software Engineering, Sun Yat-Sen University, Zhuhai, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3787-1673","authenticated-orcid":false,"given":"Dan","family":"Li","sequence":"additional","affiliation":[{"name":"School of Software Engineering, Sun Yat-Sen University, Zhuhai, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-2464-4102","authenticated-orcid":false,"given":"Jiahui","family":"Zhou","sequence":"additional","affiliation":[{"name":"School of Software Engineering, Sun Yat-Sen University, Zhuhai, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-1921-2767","authenticated-orcid":false,"given":"Shunyu","family":"Wu","sequence":"additional","affiliation":[{"name":"School of Software Engineering, Sun Yat-Sen University, Zhuhai, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-9927-8911","authenticated-orcid":false,"given":"Haozheng","family":"Ye","sequence":"additional","affiliation":[{"name":"School of Software Engineering, Sun Yat-Sen University, Zhuhai, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4110-2068","authenticated-orcid":false,"given":"Jian","family":"Lou","sequence":"additional","affiliation":[{"name":"School of Software Engineering, Sun Yat-Sen University, Zhuhai, Guangdong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6565-7511","authenticated-orcid":false,"given":"See-Kiong","family":"Ng","sequence":"additional","affiliation":[{"name":"Institute of Data Science, National University of Singapore, Singapore, Singapore and School of Computing, National University of Singapore, Singapore, Singapore"}]}],"member":"320","published-online":{"date-parts":[[2025,11,10]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3703447"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3533382"},{"key":"e_1_3_2_1_3_1","volume-title":"International Conference on Machine Learning. PMLR, 2397-2430","author":"Biderman Stella","year":"2023","unstructured":"Stella Biderman, Hailey Schoelkopf, Quentin Gregory Anthony, Herbie Bradley, Kyle O'Brien, Eric Hallahan, Mohammad Aflah Khan, Shivanshu Purohit, USVSN Sai Prashanth, Edward Raff, et al., 2023. Pythia: A suite for analyzing large language models across training and scaling. In International Conference on Machine Learning. PMLR, 2397-2430."},{"key":"e_1_3_2_1_4_1","unstructured":"Tom Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared D Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et al. 2020. Language models are few-shot learners. Advances in neural information processing systems Vol. 33 (2020) 1877-1901."},{"key":"e_1_3_2_1_5_1","volume-title":"Forty-first International Conference on Machine Learning.","author":"Das Abhimanyu","year":"2024","unstructured":"Abhimanyu Das, Weihao Kong, Rajat Sen, and Yichen Zhou. 2024. A decoder-only foundation model for time-series forecasting. In Forty-first International Conference on Machine Learning."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11831-021-09616-4"},{"key":"e_1_3_2_1_7_1","volume-title":"Moment: A family of open time-series foundation models. arXiv preprint arXiv:2402.03885","author":"Goswami Mononito","year":"2024","unstructured":"Mononito Goswami, Konrad Szafer, Arjun Choudhry, Yifu Cai, Shuo Li, and Artur Dubrawski. 2024. Moment: A family of open time-series foundation models. arXiv preprint arXiv:2402.03885 (2024)."},{"key":"e_1_3_2_1_8_1","volume-title":"Wilson","author":"Shikai Qiu Gruver Marc Finzi","year":"2023","unstructured":"Marc Finzi Shikai Qiu Gruver, Nate and Andrew G. Wilson. 2023. Large language models are zero-shot time series forecasters. Advances in Neural Information Processing Systems, 36 (2023)."},{"key":"e_1_3_2_1_9_1","volume-title":"TimeFilter: Patch-specific spatial-temporal graph filtration for time series forecasting. arXiv preprint arXiv:2501.13041","author":"Hu Yifan","year":"2025","unstructured":"Yifan Hu, Guibin Zhang, Peiyuan Liu, Disen Lan, Naiqi Li, Dawei Cheng, Tao Dai, Shu-Tao Xia, and Shirui Pan. 2025. TimeFilter: Patch-specific spatial-temporal graph filtration for time series forecasting. arXiv preprint arXiv:2501.13041 (2025)."},{"key":"e_1_3_2_1_10_1","unstructured":"Yue Jiang Yile Chen Xiucheng Li Qin Chao Shuai Liu and Gao Cong. [n.d.]. FSTLLM: Spatio-Temporal LLM for Few Shot Time Series Forecasting. In Forty-second International Conference on Machine Learning."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00324"},{"key":"e_1_3_2_1_12_1","volume-title":"Time-llm: Time series forecasting by reprogramming large language models. arXiv preprint arXiv:2310.01728","author":"Jin Ming","year":"2023","unstructured":"Ming Jin, Shiyu Wang, Lintao Ma, Zhixuan Chu, James Y Zhang, Xiaoming Shi, Pin-Yu Chen, Yuxuan Liang, Yuan-Fang Li, Shirui Pan, et al., 2023. Time-llm: Time series forecasting by reprogramming large language models. arXiv preprint arXiv:2310.01728 (2023)."},{"key":"e_1_3_2_1_13_1","volume-title":"Prefix-tuning: Optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190","author":"Li Xiang Lisa","year":"2021","unstructured":"Xiang Lisa Li and Percy Liang. 2021. Prefix-tuning: Optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190 (2021)."},{"key":"e_1_3_2_1_14_1","volume-title":"Efficient multivariate time series forecasting via calibrated language models with privileged knowledge distillation. arXiv preprint arXiv:2505.02138","author":"Liu Chenxi","year":"2025","unstructured":"Chenxi Liu, Hao Miao, Qianxiong Xu, Shaowen Zhou, Cheng Long, Yan Zhao, Ziyue Li, and Rui Zhao. 2025a. Efficient multivariate time series forecasting via calibrated language models with privileged knowledge distillation. arXiv preprint arXiv:2505.02138 (2025)."},{"key":"e_1_3_2_1_15_1","volume-title":"Lstprompt: Large language models as zero-shot time series forecasters by long-short-term prompting. arXiv preprint arXiv:2402.16132","author":"Liu Haoxin","year":"2024","unstructured":"Haoxin Liu, Zhiyuan Zhao, Jindong Wang, Harshavardhan Kamarthi, and B Aditya Prakash. 2024b. Lstprompt: Large language models as zero-shot time series forecasters by long-short-term prompting. arXiv preprint arXiv:2402.16132 (2024)."},{"key":"e_1_3_2_1_16_1","volume-title":"Timebridge: Non-stationarity matters for long-term time series forecasting. arXiv preprint arXiv:2410.04442","author":"Liu Peiyuan","year":"2024","unstructured":"Peiyuan Liu, Beiliang Wu, Yifan Hu, Naiqi Li, Tao Dai, Jigang Bao, and Shu-tao Xia. 2024a. Timebridge: Non-stationarity matters for long-term time series forecasting. arXiv preprint arXiv:2410.04442 (2024)."},{"key":"e_1_3_2_1_17_1","volume-title":"prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM computing surveys","author":"Liu Pengfei","year":"2023","unstructured":"Pengfei Liu, Weizhe Yuan, Jinlan Fu, Zhengbao Jiang, Hiroaki Hayashi, and Graham Neubig. 2023. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM computing surveys, Vol. 55, 9 (2023), 1-35."},{"key":"e_1_3_2_1_18_1","volume-title":"Zhengxiao Du, Zhilin Yang, and Jie Tang.","author":"Liu Xiao","year":"2021","unstructured":"Xiao Liu, Kaixuan Ji, Yicheng Fu, Weng Lam Tam, Zhengxiao Du, Zhilin Yang, and Jie Tang. 2021. P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks. arXiv preprint arXiv:2110.07602 (2021)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.aiopen.2023.08.012"},{"key":"e_1_3_2_1_20_1","volume-title":"Sundial: A family of highly capable time series foundation models. arXiv preprint arXiv:2502.00816","author":"Liu Yong","year":"2025","unstructured":"Yong Liu, Guo Qin, Zhiyuan Shi, Zhi Chen, Caiyin Yang, Xiangdong Huang, Jianmin Wang, and Mingsheng Long. 2025b. Sundial: A family of highly capable time series foundation models. arXiv preprint arXiv:2502.00816 (2025)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3711896.3737137"},{"key":"e_1_3_2_1_22_1","volume-title":"Non-stationary transformers: Exploring the stationarity in time series forecasting. Advances in neural information processing systems","author":"Liu Yong","year":"2022","unstructured":"Yong Liu, Haixu Wu, Jianmin Wang, and Mingsheng Long. 2022. Non-stationary transformers: Exploring the stationarity in time series forecasting. Advances in neural information processing systems, Vol. 35 (2022), 9881-9893."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3711896.3737157"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.3390\/en17071662"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3711896.3737123"},{"key":"e_1_3_2_1_26_1","volume-title":"Language models as knowledge bases? arXiv preprint arXiv:1909.01066","author":"Petroni Fabio","year":"2019","unstructured":"Fabio Petroni, Tim Rockt\u00e4schel, Patrick Lewis, Anton Bakhtin, Yuxiang Wu, Alexander H Miller, and Sebastian Riedel. 2019. Language models as knowledge bases? arXiv preprint arXiv:1909.01066 (2019)."},{"key":"e_1_3_2_1_27_1","volume-title":"Exploiting cloze questions for few shot text classification and natural language inference. arXiv preprint arXiv:2001.07676","author":"Schick Timo","year":"2020","unstructured":"Timo Schick and Hinrich Sch\u00fctze. 2020. Exploiting cloze questions for few shot text classification and natural language inference. arXiv preprint arXiv:2001.07676 (2020)."},{"key":"e_1_3_2_1_28_1","volume-title":"Mehmet Ugur Gudelek, and Ahmet Murat Ozbayoglu","author":"Sezer Omer Berat","year":"2020","unstructured":"Omer Berat Sezer, Mehmet Ugur Gudelek, and Ahmet Murat Ozbayoglu. 2020. Financial time series forecasting with deep learning: A systematic literature review: 2005-2019. Applied soft computing, Vol. 90 (2020), 106181."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3711896.3736860"},{"key":"e_1_3_2_1_30_1","volume-title":"Eric Wallace, and Sameer Singh.","author":"Shin Taylor","year":"2020","unstructured":"Taylor Shin, Yasaman Razeghi, Robert L Logan IV, Eric Wallace, and Sameer Singh. 2020. Autoprompt: Eliciting knowledge from language models with automatically generated prompts. arXiv preprint arXiv:2010.15980 (2020)."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3711896.3737005"},{"key":"e_1_3_2_1_32_1","unstructured":"Gerald Woo Chenghao Liu Akshat Kumar Caiming Xiong Silvio Savarese and Doyen Sahoo. 2024. Unified training of universal time series forecasting transformers. (2024)."},{"key":"e_1_3_2_1_33_1","volume-title":"Etsformer: Exponential smoothing transformers for time-series forecasting. arXiv preprint arXiv:2202.01381","author":"Woo Gerald","year":"2022","unstructured":"Gerald Woo, Chenghao Liu, Doyen Sahoo, Akshat Kumar, and Steven Hoi. 2022. Etsformer: Exponential smoothing transformers for time-series forecasting. arXiv preprint arXiv:2202.01381 (2022)."},{"key":"e_1_3_2_1_34_1","volume-title":"Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv preprint arXiv:2210.02186","author":"Wu Haixu","year":"2022","unstructured":"Haixu Wu, Tengge Hu, Yong Liu, Hang Zhou, Jianmin Wang, and Mingsheng Long. 2022. Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv preprint arXiv:2210.02186 (2022)."},{"key":"e_1_3_2_1_35_1","volume-title":"Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in neural information processing systems","author":"Wu Haixu","year":"2021","unstructured":"Haixu Wu, Jiehui Xu, Jianmin Wang, and Mingsheng Long. 2021. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in neural information processing systems, Vol. 34 (2021), 22419-22430."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2023.3342137"},{"key":"e_1_3_2_1_37_1","volume-title":"Less Is More: Fast Multivariate Time Series Forecasting with Light Sampling-oriented MLP Structures. arXiv. arXiv preprint arXiv:2207.01186","author":"Zhang T","year":"2022","unstructured":"T Zhang, Y Zhang, W Cao, J Bian, X Yi, S Zheng, and J Li. 2022. Less Is More: Fast Multivariate Time Series Forecasting with Light Sampling-oriented MLP Structures. arXiv. arXiv preprint arXiv:2207.01186 (2022)."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3711896.3736872"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"e_1_3_2_1_40_1","volume-title":"International conference on machine learning. PMLR, 27268-27286","author":"Zhou Tian","year":"2022","unstructured":"Tian Zhou, Ziqing Ma, Qingsong Wen, Xue Wang, Liang Sun, and Rong Jin. 2022. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In International conference on machine learning. PMLR, 27268-27286."}],"event":{"name":"CIKM '25: The 34th ACM International Conference on Information and Knowledge Management","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval","SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Seoul Republic of Korea","acronym":"CIKM '25"},"container-title":["Proceedings of the 34th ACM International Conference on Information and Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3746252.3760803","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T00:41:45Z","timestamp":1765500105000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3746252.3760803"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,10]]},"references-count":40,"alternative-id":["10.1145\/3746252.3760803","10.1145\/3746252"],"URL":"https:\/\/doi.org\/10.1145\/3746252.3760803","relation":{},"subject":[],"published":{"date-parts":[[2025,11,10]]},"assertion":[{"value":"2025-11-10","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}