{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T00:46:59Z","timestamp":1765500419196,"version":"3.48.0"},"publisher-location":"New York, NY, USA","reference-count":32,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,11,10]]},"DOI":"10.1145\/3746252.3760828","type":"proceedings-article","created":{"date-parts":[[2025,11,10]],"date-time":"2025-11-10T18:37:32Z","timestamp":1762799852000},"page":"5238-5242","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["LLM4ES: Learning User Embeddings from Event Sequences via Large Language Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-5207-7610","authenticated-orcid":false,"given":"Aleksei","family":"Shestov","sequence":"first","affiliation":[{"name":"Sber AI Lab, Moscow, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-9447-5599","authenticated-orcid":false,"given":"Omar","family":"Zoloev","sequence":"additional","affiliation":[{"name":"Sber AI Lab, Moscow, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6761-8695","authenticated-orcid":false,"given":"Maksim","family":"Makarenko","sequence":"additional","affiliation":[{"name":"Sber AI Lab, Moscow, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-1978-9995","authenticated-orcid":false,"given":"Mikhail","family":"Orlov","sequence":"additional","affiliation":[{"name":"Sber AI Lab, Moscow, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-2111-6760","authenticated-orcid":false,"given":"Egor","family":"Fadeev","sequence":"additional","affiliation":[{"name":"Sber AI Lab, Moscow, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-1618-8981","authenticated-orcid":false,"given":"Ivan","family":"Kireev","sequence":"additional","affiliation":[{"name":"Sber AI Lab, Moscow, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6196-0564","authenticated-orcid":false,"given":"Andrey","family":"Savchenko","sequence":"additional","affiliation":[{"name":"Sber AI Lab, Moscow, Russian Federation and Laboratory of Algorithms and Technologies for Networks Analysis, HSE University, Nizhny Novgorod, Russian Federation"}]}],"member":"320","published-online":{"date-parts":[[2025,11,10]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"ESQA: Event Sequences Question Answering. arXiv preprint arXiv:2407.12833","author":"Abdullaeva Irina","year":"2024","unstructured":"Irina Abdullaeva, Andrei Filatov, Mikhail Orlov, Ivan Karpukhin, Viacheslav Vasilev, Denis Dimitrov, Andrey Kuznetsov, Ivan Kireev, and Andrey Savchenko. 2024. ESQA: Event Sequences Question Answering. arXiv preprint arXiv:2407.12833 (2024)."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3514221.3526129"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330693"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3608857"},{"key":"e_1_3_2_1_5_1","volume-title":"First Conference on Language Modeling.","author":"BehnamGhader Parishad","year":"2024","unstructured":"Parishad BehnamGhader, Vaibhav Adlakha, Marius Mosbach, Dzmitry Bahdanau, Nicolas Chapados, and Siva Reddy. 2024. LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders. In First Conference on Language Modeling."},{"key":"e_1_3_2_1_6_1","volume-title":"Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555","author":"Clark Kevin","year":"2020","unstructured":"Kevin Clark, Minh-Thang Luong, Quoc V Le, and Christopher D Manning. 2020. Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555 (2020)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599287"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599534"},{"key":"e_1_3_2_1_9_1","first-page":"4171","volume-title":"Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies","volume":"1","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers). 4171-4186."},{"key":"e_1_3_2_1_10_1","unstructured":"Abhimanyu Dubey Abhinav Jauhri Abhinav Pandey Abhishek Kadian Ahmad Al-Dahle Aiesha Letman Akhil Mathur Alan Schelten Amy Yang Angela Fan et al. 2024. The llama 3 herd of models. arXiv preprint arXiv:2407.21783 (2024)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599803"},{"key":"e_1_3_2_1_12_1","volume-title":"A survey on self-supervised learning: Algorithms, applications, and future trends","author":"Gui Jie","year":"2024","unstructured":"Jie Gui, Tuo Chen, Jing Zhang, Qiong Cao, Zhenan Sun, Hao Luo, and Dacheng Tao. 2024. A survey on self-supervised learning: Algorithms, applications, and future trends. IEEE Transactions on Pattern Analysis and Machine Intelligence (2024)."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/2827872"},{"key":"e_1_3_2_1_14_1","volume-title":"International Conference on Artificial Intelligence and Statistics. PMLR, 5549-5581","author":"Hegselmann Stefan","year":"2023","unstructured":"Stefan Hegselmann, Alejandro Buendia, Hunter Lang, Monica Agrawal, Xiaoyi Jiang, and David Sontag. 2023. Tabllm: Few-shot classification of tabular data with large language models. In International Conference on Artificial Intelligence and Statistics. PMLR, 5549-5581."},{"key":"e_1_3_2_1_15_1","volume-title":"LoRA: Low-Rank Adaptation of Large Language Models. In International Conference on Learning Representations.","author":"Hu Edward J","year":"2022","unstructured":"Edward J Hu, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al., 2022. LoRA: Low-Rank Adaptation of Large Language Models. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_16_1","volume-title":"The Twelfth International Conference on Learning Representations.","author":"Jin Ming","year":"2024","unstructured":"Ming Jin, Shiyu Wang, Lintao Ma, Zhixuan Chu, James Y Zhang, Xiaoming Shi, Pin-Yu Chen, Yuxuan Liang, Yuan-Fang Li, Shirui Pan, et al., 2024. Time-LLM: Time Series Forecasting by Reprogramming Large Language Models. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41597-023-01995-1"},{"key":"e_1_3_2_1_18_1","volume-title":"DeTPP: Leveraging Object Detection for Robust Long-Horizon Event Prediction. arXiv preprint arXiv:2408.13131","author":"Karpukhin Ivan","year":"2024","unstructured":"Ivan Karpukhin and Andrey Savchenko. 2024. DeTPP: Leveraging Object Detection for Robust Long-Horizon Event Prediction. arXiv preprint arXiv:2408.13131 (2024)."},{"key":"e_1_3_2_1_19_1","first-page":"3149","volume-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems","author":"Ke Guolin","year":"2017","unstructured":"Guolin Ke, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu. 2017. LightGBM: a highly efficient gradient boosting decision tree. In Proceedings of the 31st International Conference on Neural Information Processing Systems (Long Beach, California, USA) (NIPS'17). Curran Associates Inc., Red Hook, NY, USA, 3149-3157."},{"key":"e_1_3_2_1_20_1","volume-title":"Proceedings of the 11th International Conference on Learning Representations (ICLR","author":"Liu Zitao","year":"2023","unstructured":"Zitao Liu, Qiongqiong Liu, Jiahao Chen, Shuyan Huang, and Weiqi Luo. 2023a. simpleKT: A Simple But Tough-to-Beat Baseline for Knowledge Tracing. In Proceedings of the 11th International Conference on Learning Representations (ICLR 2023). https:\/\/openreview.net\/forum?id=S0EXpZJ3sXM"},{"key":"e_1_3_2_1_21_1","volume-title":"Advances in Neural Information Processing Systems 36: Datasets and Benchmarks Track (NeurIPS","author":"Liu Zitao","year":"2023","unstructured":"Zitao Liu, Qiongqiong Liu, Teng Guo, Jiahao Chen, Shuyan Huang, Xiangyu Zhao, Jiliang Tang, Weiqi Luo, and Jian Weng. 2023b. XES3G5M: A Knowledge Tracing Benchmark Dataset with Auxiliary Information. In Advances in Neural Information Processing Systems 36: Datasets and Benchmarks Track (NeurIPS 2023). https:\/\/datasets-benchmarks-proceedings.neurips.cc\/paper\/2023\/file\/67fc628f17c2ad53621fb961c6bafcaf-Paper.pdf"},{"key":"e_1_3_2_1_22_1","volume-title":"Meta AI Blog. Retrieved December","volume":"20","author":"Meta AI","year":"2024","unstructured":"AI Meta. 2024. Llama 3.2: Revolutionizing edge AI and vision with open, customizable models. Meta AI Blog. Retrieved December, Vol. 20 (2024), 2024."},{"key":"e_1_3_2_1_23_1","volume-title":"Generative Representational Instruction Tuning. In ICLR 2024 Workshop: How Far Are We From AGI.","author":"Muennighoff Niklas","year":"2024","unstructured":"Niklas Muennighoff, Hongjin SU, Liang Wang, Nan Yang, Furu Wei, Tao Yu, Amanpreet Singh, and Douwe Kiela. 2024. Generative Representational Instruction Tuning. In ICLR 2024 Workshop: How Far Are We From AGI."},{"key":"e_1_3_2_1_24_1","volume-title":"Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748","author":"van den Oord Aaron","year":"2018","unstructured":"Aaron van den Oord, Yazhe Li, and Oriol Vinyals. 2018. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3608783"},{"key":"e_1_3_2_1_26_1","volume-title":"Intensity-Free Learning of Temporal Point Processes. In International Conference on Learning Representations.","author":"Shchur Oleksandr","year":"2020","unstructured":"Oleksandr Shchur, Marin Bilo\u0161, and Stephan G\u00fcnnemann. 2020. Intensity-Free Learning of Temporal Point Processes. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3604237.3626850"},{"key":"e_1_3_2_1_28_1","volume-title":"The Twelfth International Conference on Learning Representations.","author":"Sun Chenxi","year":"2024","unstructured":"Chenxi Sun, Hongyan Li, Yaliang Li, and Shenda Hong. 2024. TEST: Text Prototype Aligned Embedding to Activate LLM's Ability for Time Series. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_2_1_29_1","volume-title":"Pooling and attention: What are effective designs for llm-based embedding models? arXiv preprint arXiv:2409.02727","author":"Tang Yixuan","year":"2024","unstructured":"Yixuan Tang and Yi Yang. 2024. Pooling and attention: What are effective designs for llm-based embedding models? arXiv preprint arXiv:2409.02727 (2024)."},{"key":"e_1_3_2_1_30_1","unstructured":"Qwen Team. 2025. Qwen3 Technical Report. arXiv:2505.09388 [cs.CL] https:\/\/arxiv.org\/abs\/2505.09388"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3608874"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3589334.3645358"}],"event":{"name":"CIKM '25: The 34th ACM International Conference on Information and Knowledge Management","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval","SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Seoul Republic of Korea","acronym":"CIKM '25"},"container-title":["Proceedings of the 34th ACM International Conference on Information and Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3746252.3760828","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,12]],"date-time":"2025-12-12T00:43:42Z","timestamp":1765500222000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3746252.3760828"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,10]]},"references-count":32,"alternative-id":["10.1145\/3746252.3760828","10.1145\/3746252"],"URL":"https:\/\/doi.org\/10.1145\/3746252.3760828","relation":{},"subject":[],"published":{"date-parts":[[2025,11,10]]},"assertion":[{"value":"2025-11-10","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}